If You Only Read Ten Papers, Read These
The Generative Engine Optimization literature has grown to several dozen papers. Most practitioners have read zero of them. That is understandable; academic papers are long, dense, and often separated from practitioner needs by jargon. But a handful of studies genuinely change how you should think about AI visibility. This guide is an editorial selection of the ten most important papers for practitioners, ranked by how much each paper will change the decisions you make on Monday morning. Each entry summarizes the finding in plain language, explains why it matters for your brand, and points you to the specific practice it should change.
Our larger GEO academic papers synthesis catalogues 30+ papers; this guide is the curated subset we recommend to every new Presenc AI customer as required reading before their first strategy session.
1. Aggarwal et al., "GEO: Generative Engine Optimization" (Princeton, 2024)
The finding: Adding citations, quotations, and statistics to content produces a 40% lift in visibility metrics across major generative engines.
Why you need to read it: This is the paper that founded the field. Everything else in GEO builds on the Princeton team's framing of visibility metrics (Position-adjusted Word Count, Subjective Impression) and their GEO-bench dataset. Read it to understand the vocabulary the rest of the field uses.
What to change Monday: Audit your highest-traffic content for citation, quote, and statistic density. Add three of each to every canonical page over the next sprint.
2. Chen et al., "What Generative Search Engines Like" (CMU, 2025)
The finding: Cooperative, semantically-explicit content earns 35–60% higher citation rate than adversarial or terse equivalents. AI engines cite content that makes their job easier.
Why you need to read it: Shifts the frame from "trick the AI" to "help the AI." Validated across multiple engines with controlled experiments.
What to change Monday: Rewrite thesis sentences at the top of sections to be crisp, standalone summaries. Make it easy for the AI to extract your answer.
3. Kumar + Palkhouski, "GEO-16 Framework for B2B SaaS" (UC Berkeley, 2025)
The finding: 16-pillar auditing framework with novel signals including answer density, entity tag density, and quote-to-prose ratio. Applied to B2B SaaS with empirical validation across 70 industry prompts.
Why you need to read it: Most GEO audits miss half of what matters. The GEO-16 framework gives you a checklist that is both comprehensive and empirically validated.
What to change Monday: Score your top 20 pages against the 16 pillars. The lowest three scores usually point to your highest-ROI fixes.
4. Schanbacher, "The Impact of JSON-LD Metadata on ChatGPT Visibility" (Hochschule Furtwangen, 2025)
The finding: Empirical study of real estate agencies showing that sites with rich JSON-LD markup are significantly more likely to be "known" to ChatGPT than sites without.
Why you need to read it: Makes the schema-markup case with data, not assertion. If you have ever needed to justify a schema-coverage project to non-technical stakeholders, this paper is the citation.
What to change Monday: Complete Organization, Article, and Product schema across every canonical page. This is the single cheapest measurable AI-visibility lift available.
5. Cook, "Mind the Invisibility Gap: 80 Global Brands in LLM Answers" (Geometriqs, 2025)
The finding: Even Fortune-500-equivalent brands are systematically underrepresented in LLM answers relative to market share. Cross-model consistency is lower than typically assumed.
Why you need to read it: Disabuses enterprise stakeholders of the notion that brand size alone guarantees AI visibility. Great for internal "why should we invest in GEO" conversations.
What to change Monday: Benchmark your brand against the Global 80 results. If a Fortune 500 is underweighted in AI responses, your mid-market brand is likely much worse off.
6. Chen et al., "Role-Augmented Intent-Driven GEO" (USTC, 2025)
The finding: Tailoring content to specific personas (persona + role + intent) before LLM ingestion improves citation rates 11–23% over generic optimization.
Why you need to read it: Single-voice pages lose to persona-matched pages. If you sell to multiple buyer personas, this has direct implications for your content architecture.
What to change Monday: Audit your top pages for persona clarity. If a single page tries to address CFOs, developers, and CMOs, split it into three pages.
7. Anthropic, "A small number of samples can poison LLMs of any size" (2025)
The finding: 250 malicious documents can install backdoors in LLMs regardless of model size. Disproves the assumption that attackers need percentage-scale corpus control.
Why you need to read it: AI brand safety is cheap to attack. This changes how seriously to take adversarial monitoring for regulated industries and contested brand categories.
What to change Monday: Include adversarial content monitoring in your brand-safety dashboards. Coordinated smear campaigns in the corpora that feed major LLMs (Reddit threads, indexed PDFs) are now a documented attack surface.
8. Chen, Wang, Chen, Koudas, "Navigating the Shift" (University of Toronto, 2026)
The finding: AI responses systematically favor certain source types (Wikipedia, news, government) over traditional search, which shows greater source diversity.
Why you need to read it: If your brand is not on Wikipedia or in major news, your AI visibility has a structural ceiling. This paper quantifies how much.
What to change Monday: Audit your Wikipedia entry and Wikidata record. If either is thin or missing, prioritize earning a compliant entry or update this quarter.
9. Sharma, "The Discovery Gap: Product Hunt Startups in LLM Discovery" (IIT Patna, 2025)
The finding: Of 112 Product Hunt top-500 startups, only ~12% surfaced in ChatGPT discovery queries. Product Hunt ranking had near-zero correlation with LLM visibility.
Why you need to read it: Founders often assume a great Product Hunt launch translates into discoverability. This paper shows it does not. Discoverability is a separate investment.
What to change Monday: If you are early-stage, stop relying on launch moments for discovery. Budget dedicated GEO work from day 1.
10. Kumar + Lakkaraju, "Manipulating LLMs to Increase Product Visibility" (Harvard, 2024)
The finding: Adversarial prompts can reliably increase LLM recommendation rates for targeted products. Detection is difficult.
Why you need to read it: Defines the ethical boundary. Helps you recognize when a "GEO vendor" crosses into adversarial tactics that create legal and brand-safety risk.
What to change Monday: Audit any third-party GEO vendor's methodology. If they cannot explain their tactics in terms of content-quality signals, be cautious.
How to Use These Papers Inside Your Team
The most effective way to operationalize these papers is to pick one per week and have someone read it plus share a one-page practitioner summary with your team. In 10 weeks your marketing team will have a shared vocabulary and empirical grounding that almost no competitor has. This is the fastest path to a serious internal GEO capability.
For leadership briefings, we recommend the Cook Geometriqs Global 80 paper as the single most persuasive "why should we fund this" reference. It documents that even top global brands are underweighted in AI responses, which disarms the "our brand is too big to need this" objection.
Beyond These Ten
Once you have read these ten, our full GEO academic papers synthesis covers 30+ papers including measurement frameworks (AIVO Standard), security and poisoning studies, and industry-specific investigations. For hands-on implementation, pair the research with our AI visibility audit guide and the audit checklist template.