What the 2025 GEO Study Reveals About Ranking in AI Answers (and How I Apply It)

TL;DR

A 2025 research study from researchers at MIT and Columbia University tested what actually moves content up in AI-generated rankings. Their big finding wasn’t “use this one weird prompt.” It was that repeatable, reliable gains come from a consistent structure: match user intent, stay factual, show clear differentiation, use evidence, and format content so it’s easy to extract.

I built a custom internal tool inspired by that study to help me (and my clients) turn messy, long-form information into “AI-ready” answer blocks that are more likely to be pulled and preferred in AI results. We’re actively expanding that tool to make it more comprehensive, and it’s already being used to guide client content.

I’ve spent the last couple of years watching the same pattern play out across industries:

  • Some brands get mentioned in AI answers… but show up as the third or fourth recommendation.
  • Others get pulled sometimes… but inconsistently.
  • And a few dominate because their content is the easiest to trust, quote, and rank.

That’s why I pay attention when researchers publish something that goes beyond opinions and actually tests what works.

The research team from MIT and Columbia University ran the study (E-GEO: A Testbed for Generative Engine Optimization in E-Commerce) to measure how answer engine rankings change based on how content is written. They used thousands of real-world shopping-style queries, rewrote product descriptions using different approaches, and tracked whether the rewritten versions moved up or down in AI rankings.

What I’m sharing below is the practical version of what matters, and how you can apply it whether you sell products or services.

The two-step reality: getting pulled vs getting preferred

RAG process
Summary of RAG process

Most brands only think about one part of the problem.

Step 1: Retrieval (getting pulled in)

This is whether your page, brand, or product even gets included as a possible answer.

Step 2: Re-ranking (getting preferred)

This is where the AI decides what’s “best,” “second best,” and so on.

The 2025 study focused heavily on that second step: once options are in the set, what makes one rise to the top?

If you’re investing in content for AI visibility, you want both:

● content that reliably gets pulled
● content that reliably gets ranked as the preferred option

What the researchers actually did (in plain English)

Reddit queries matched to Amazon Listings
Reddit queries matched to Amazon Listings

Here’s the simple version of the test:

  1. They took 7,000+ real “what should I buy?” style posts from Reddit.
  2. They matched each query to the 10 most relevant Amazon product listings using a semantic similarity method (basically, “meaning match,” not just keyword match).
  3. They rewrote product descriptions using different prompt styles.
  4. They measured whether the rewritten descriptions moved up in the AI’s ranking.
  5. They then used a second model to iteratively improve the rewrite prompts until the rewrites performed better.

The purpose wasn’t to find a trick. It was to find patterns that hold up repeatedly across many queries and products.

That’s exactly what brands need right now, repeatable rules, not hype.

The most important practical takeaway

ranking in ChatGPT
ranking in ChatGPT

AI rankings reward content that makes it easy to confidently choose and easy to quote.

That’s why the winning patterns consistently included:

1) Clear intent match

The best performing rewrites aligned tightly with what the user actually asked for, especially long, conversational queries with constraints.

Not “Knife set for kitchen.”

More like:
“Premium durable knife set with minimal upkeep needed.”

That shift matters because it directly mirrors the user’s real goal.

2) Factual grounding

One of the most consistent themes was factuality. Keep claims accurate, avoid embellishment, and preserve what can be supported.

In real life, that means:

● don’t guess
● don’t inflate benefits
● don’t claim “best” without support

3) Clear differentiation (your “why us” without fluff)

Competitive positioning mattered, especially when it was expressed as concrete differences, not generic marketing language.

Examples:

● materials, specs, certifications
● what’s included vs not included
● warranty terms
● durability, maintenance requirements
● constraints the product/service is best for (and not best for)

4) Evidence signals

When you can back something up with data or reputable references, do it. Evidence helps the AI system “trust” the content and reuse it.

For service businesses, evidence can be:

● licensing and certifications
● documented process steps
● before/after metrics
● review volume and rating
● pricing ranges and what drives them

5) Scannable formatting

This one is huge, and it matches what I’ve seen in the wild. Content that’s easy to scan is easier for AI systems to lift into answers.

Headings, bullets, short blocks, ranges, and direct definitions beat long paragraphs every time.

What didn’t work as well

GEO and AEO Heuristics that did not work well
GEO and AEO Heuristics that did not work well

A lot of “internet advice” about AI visibility leans into tone or style:

● “Write like an ad”
● “Be super persuasive”
● “Tell a story”
● “Sound authoritative”

The study showed that these approaches can be inconsistent, and in some cases can even hurt performance if they reduce factual clarity or drift away from the user’s intent.

The takeaway: style is secondary; structure and usefulness come first.

The custom tool I built (inspired by the study)

meta optimizer for GEO and AEO
meta optimizer for GEO and AEO

This study didn’t just confirm what I suspected, it gave me a structure I could build around.

So I built an internal GPT-powered re-ranking workflow inspired by the study’s re-ranking logic.

Here’s what it does in practical terms:

What my tool does today

  1. I feed it a target query (example: “How much does carpet cleaning cost in Northern Virginia?”).
  2. It ingests multiple content “chunks” pulled from top-performing pages on Google.
  3. It ranks those chunks based on which one most directly and completely answers the query.
  4. It rewrites the best chunk into multiple output formats, especially:

○ a tight “answer block”
○ a highly scannable version with headings and bullets

  1. It re-ranks the outputs and recommends the version most likely to be extractable and preferred in AI answers.

In short, it helps us consistently produce content that is accurate, aligned with intent, and formatted for extraction, which is exactly what the study suggests is repeatably effective.

How we’re expanding it

Right now we’re expanding the workflow to be more iterative, and mathematical about which content is most optimized. This will result in:

● content that more accurately reflects what AI engines reward
● stronger evidence handling
● consistency checks
● broader content formats

How we’re using it for clients

We’re already using this tool in our client work to:

● upgrade existing pages into AI-ready “answer-first” structures
● produce scannable sections that AI engines can quote cleanly
● reduce fluff while increasing proof and clarity
● align content with how people actually ask questions in AI tools

This is one of the ways we’ve been able to move faster while staying grounded and factual, because the tool forces discipline around intent, structure, and evidence.

Final thoughts

ai seo
ai seo

The 2025 MIT + Columbia E-GEO study supports something I’ve been emphasizing for a while:

If you want to rank well in AI answers, your content has to do more than “sound good.” It has to be the clearest match to the user’s intent, backed by facts, and formatted in a way that’s easy to extract and trust.

That’s why I built a tool around this, and why we’re expanding it and using it actively in client content.

If you want your site to show up more often and be preferred in AI answers, visit seorankmedia.com and request my free AI visibility checklist. I’ll review your site and tell you:

● what’s preventing AI engines from pulling your pages
● what’s keeping you from being ranked as the “top” recommendation
● which pages to fix first for the fastest impact

You can implement the changes yourself, or my team at SEO Rank Media can handle it for you.