How ChatGPT Handles Transactional Intent in Finance

BrightEdge AI Hypercube analysis reveals that a large share of finance-related ChatGPT prompts reflect transactional intent, showing that users are not only seeking information but are actively trying to take action on financial products and services.

People turn to ChatGPT with financial questions every day. But a closer look at the prompts reveals something that challenges a core assumption about how AI search works: a significant portion of finance-related ChatGPT prompts aren't informational at all. They reflect transactional intent. People aren't just learning about financial products. They're trying to act on them.

We used BrightEdge AI Hypercube to analyze finance-related prompts in ChatGPT, focusing on queries tied to the top U.S. banks and financial institutions. The patterns that emerged are consistent, significant, and directly relevant to any financial services brand thinking about AI visibility strategy.

Data Collected

 

Data PointDescription
Prompt classificationFinance-related prompts in ChatGPT filtered to transactional intent using BrightEdge AI Hypercube classification
Intent cluster analysisTransactional prompts grouped into behavioral clusters based on the action type being expressed
Volume analysisBrightEdge monthly prompt volume data applied to identify the highest-frequency transactional query patterns
Branded vs. non-branded segmentationPrompts analyzed for the presence of specific brand names to determine whether branded or generic queries drive higher volume
Agentic prompt identificationPrompts written as direct instructions to an agent isolated and analyzed as a distinct behavioral pattern

Key Finding

Finance is widely treated as an informational vertical in search strategy. The assumption is that people use AI to research, compare, and learn before taking action elsewhere. But the prompt data tells a different story. Transactional intent is present throughout ChatGPT finance queries, from credit card applications and pre-approval requests to account opening, credit access management, and full agentic requests to initiate financial processes. The implication for financial services brands is direct: an AI visibility strategy built only around educational and informational content is incomplete.

Five Transactional Intent Clusters in ChatGPT Finance Prompts

Credit card applications and shopping represent the single largest cluster of transactional finance prompts in ChatGPT, accounting for roughly 72% of identified transactional volume. Nearly 90,000 monthly prompts reflect people actively comparing or applying for credit cards. The majority are branded, tied to specific co-branded retail cards, airline loyalty products, and hotel programs. A single co-branded retail card drives approximately 66,000 monthly prompts on its own. These are not research prompts. They are decision prompts from people who have already identified what they want and are using ChatGPT to act on it.

Credit card pre-approval is a distinct sub-pattern worth separating out. We tracked 4,465 monthly prompts for pre-approval queries. Pre-approval is the first step in an application process. Someone asking ChatGPT about pre-approval is not exploring options. They are starting the process of applying.

Banking operations prompts reveal a different kind of transactional behavior. Prompts for opening accounts, setting up direct deposit, initiating bill pay, and processing transfers collectively represent approximately 18% of transactional finance volume. These are people using ChatGPT as a step-by-step guide for completing a banking action. The AI is functioning as a procedural concierge, not an information source.

Credit access management prompts, including queries about checking credit scores, unfreezing credit, and resolving credit report issues, account for approximately 4% of transactional volume. The significance here is what these prompts signal about intent state. Unfreezing a credit file is a prerequisite to submitting an application. A person asking ChatGPT how to unfreeze their credit is not casually curious. They are preparing to apply for something.

Agentic prompts represent an emerging pattern that sits apart from the volume-driven clusters. These are prompts written not as questions but as direct instructions: "Help me start a mortgage refinance application," "Find lenders offering competitive rates for a $250,000 loan," "Get me a personalized refinance payment estimate." These prompts do not have traditional keyword search equivalents. They reflect a user treating ChatGPT as an agent capable of initiating or facilitating a financial process, not just describing one. The behavior is early but directionally significant.

Intent Cluster Distribution

ClusterShare of Transactional Volume
Credit Card Applications and Shopping~72%
Banking Operations~18%
Loans and Lending~4%
Credit Score and Access Management~4%
Home Buying and Mortgage~2%

The Branded Signal

Branded prompts, those naming a specific financial institution, retail partner, or co-branded product, drive higher average monthly volume than non-branded equivalents. This is counterintuitive. Broad generic queries might be expected to dominate volume. Instead, the data shows people arriving in ChatGPT with a specific brand already in mind, using the AI to help them take action on a decision already made. Brand presence in AI-generated responses matters most at the exact moment of highest intent.

What Marketers Need to Know

Transactional intent in ChatGPT is real and present in your category right now. The assumption that AI search is a top-of-funnel channel does not hold in finance. People are comparing, pre-qualifying, and initiating applications inside ChatGPT. Content strategy and AI visibility strategy need to account for where people are in the decision process, not just what they want to learn.

AI agents need access to the parts of your site where action happens. When someone asks ChatGPT to find a credit card or start a loan application, the AI needs to be able to surface product pages, application entry points, and offer pages. If your AI-accessible content footprint consists primarily of blog posts and educational resources, you are optimizing for the wrong moment.

Branded intent is concentrated at the decision stage. People arriving in ChatGPT with a specific brand in mind are not at the beginning of their journey. They have already done their consideration work. If your brand is not visible at that moment, you are losing customers who were already sold.

Agentic prompts are an early signal of where this goes. The queries without traditional keyword volume are not low priority. They represent the leading edge of how consumers will interact with AI in high-consideration financial decisions. The brands that understand and optimize for this behavior now will own that conversation as it scales.

Technical Methodology

ParameterDetail
Data SourceBrightEdge AI Hypercube
Engine AnalyzedChatGPT
Query SetFinance-related prompts tied to top U.S. banking institutions, filtered to transactional intent classification
Intent ClassificationTransactional intent defined as prompts reflecting a user's goal to initiate, complete, or advance a financial action
Volume DataBrightEdge monthly prompt volume applied where available across identified transactional prompts
Branded ClassificationPrompts scored as branded when containing a named financial institution, retail partner, or co-branded product

Key Takeaways

FindingDetail
Finance ChatGPT prompts skew transactionalAcross clusters, the dominant behavior is action-oriented, not research-oriented
Credit decisions dominateRoughly 72% of transactional finance volume is tied to credit card applications and shopping
Branded prompts outperform generic on volumePeople with a specific brand in mind arrive in ChatGPT ready to act, not explore
Banking operations are a concierge use casePeople use ChatGPT to complete banking tasks step by step, not just to learn about them
Agentic prompts are emergingDirect-instruction prompts signal a shift toward AI-facilitated financial action, not just AI-assisted research
AI visibility strategy must go beyond top of funnelProduct pages, application entry points, and transactional content need to be AI-accessible to capture decision-stage intent

Download the Full Report

Download the full AI Search Report — How ChatGPT Handles Transactional Intent in Finance

Click the button above to download the full report in PDF format.

Published on  April 9, 2026

BrightEdge Data: AI Search is Reaching a Tipping Point – by End of 2026 Most Online Customers will be AI Agents

BrightEdge Data: AI Search is Reaching a Tipping Point – by End of 2026 Most Online Customers will be AI Agents

AI agents increasingly influence buying decisions, yet most companies lack a clear strategy for granting those agents access to their content 

 

SAN MATEO, Calif. — April 8, 2026 — BrightEdge, the global leader in AI-powered enterprise performance marketing and SEO, today released data revealing AI agent requests reached 88% of human organic search activity, a level nearly matching human search and signaling a fundamental shift in how customers find and evaluate brands online. AI agents are not just assisting users, they are actively shaping which companies customers choose to engage with and buy from. Based on current growth trends, BrightEdge projects that AI Agent activity will surpass human-driven search by the end of the year 2026.  

While most brands still optimize for human visitors, AI agents are already acting on behalf of those customers at nearly the same scale and likely without brands realizing it. Brands with outdated bot policies, inaccessible content, or poor AI visibility risk losing traffic and revenue. At key decision-making moments, AI systems are more likely to surface content that is easier for agents to access and interpret. For businesses, this means AI is quickly becoming a critical gatekeeper in how customers choose between brands. 

Agent activity already accounts for approximately 15% of total website traffic. Of that percentage, 95% was driven by OpenAI.  

“This is more than a traffic trend. It is a visibility challenge, a brand control challenge, and increasingly, a revenue challenge,” said Jim Yu, CEO of BrightEdge. “For years, brands have built their web presence for humans. Now, not only do they have to build for people, they need to build for AI agents acting on behalf of people. If you block or fail to optimize for these agents, you’re not blocking bots — you’re blocking customers. If brands do not make their digital presence accessible to AI agents, they risk invisibility at the exact moment customers are looking for them.” 

The shift is already happening, and most brands aren’t ready 

AI agents are now embedded in everyday customer behavior through platforms like ChatGPT, Perplexity, Gemini, and Claude. As agent activity accelerates, BrightEdge warns that many organizations are underestimating both the scale of the shift and their lack of preparedness. 

AI agents are already interacting with brand websites in significant numbers, but because this activity does not appear in traditional analytics platforms, like Google Analytics, most companies have no visibility – leaving them on the sidelines as AI agents influence which brands get recommended, and how. 

This lack of visibility translates directly into a lack of strategy. While customers are already using AI agents to research and make decisions, companies have not defined how they want those agents to access, interpret, or represent their content. BrightEdge analysis shows that only 19% of sites have specific directives for ChatGPT-related bots, and the policies vary widely. The other 81% of companies currently treat AI agents like traditional bots, applying obsolete or conflicting rules and limiting how effectively their content can be surfaced by AI systems. 

Unlike traditional bots, AI agents can influence how brands appear in customer journeys in two important ways: 

Real-time retrieval agents access websites on behalf of users and pull current information such as product details, pricing, specifications, and other decision-making content. In these moments, brands often have a single opportunity to influence how they are represented in AI-generated recommendations.

  • Training agents help shape how AI models understand brands over time, influencing how companies, products, and value propositions may be described in future AI-generated responses. This shapes not just visibility, but how a brand is positioned against competitors in AI-driven experiences. 

Together, these agents are redefining digital discovery, shifting it from direct website visits to AI-mediated decision-making. BrightEdge found that most companies focus on blocking training agents (77%), while far fewer address search (21%) or user-facing agents (38%), highlighting a lack of clear, consistent strategy for managing AI agents. 

If brands fail to act, the result can be significant: 

Brands may become less visible in AI-driven search and discovery experiences.

  • Competitors may shape the narrative and win the revenue if their content is more accessible to AI agents at critical decision-making moments.
  • Customers may receive outdated or incomplete information about products and services. 

Even in an optimistic scenario where 80% of companies correctly manage website policies for agent traffic, the remaining 20% would still translate into $40 billion in unoptimized search opportunity across the broader search economy, based on BrightEdge modeling. 

A new cross-functional priority for marketing and IT 

BrightEdge says the solution is not simply to allow all bots, but to create a more deliberate strategy for which agents brands want to welcome, what content they want surfaced, and how their sites support AI retrieval and representation. 

This marks a new operational shift: managing AI agent access is no longer just an SEO decision. It is a shared responsibility across marketing, IT, and digital teams. 

This requires coordination across teams: 

CMOs need visibility into how AI agents may affect customer discovery and brand presence. 

  • CIOs and technical teams need to revisit bot policies and site access rules.
  • Digital marketing and SEO leaders need to ensure brand, product, and conversion-critical content is accessible, accurate, and current. 

Brands that proactively enable and optimize for these systems will gain a structural advantage, while those that delay risk becoming invisible in the next generation of search. 

To help organizations better understand and act on this shift, BrightEdge recently introduced AI Hyper Cube and AI Agent Insights, two purpose-built capabilities that together provide a complete view of how brands show up across AI-driven search and agent interactions. AI Hyper Cube reveals how brands are represented in AI-generated experiences, while AI Agent Insights surfaces which agents are accessing a site and what content they engage with. Together, these capabilities enable marketing and digital teams to move from blind spots to actionable insight as AI agents increasingly shape customer discovery. For more information, visit BrightEdge.com 

About BrightEdge 

BrightEdge is the global leader in Enterprise SEO and AI-powered content performance. For more than 18 years, BrightEdge has helped thousands of brands and digital marketers, including 57% of the Fortune 500, transform online opportunities into measurable business results. Its industry-first platform integrates the most comprehensive dataset in search, combining insights from traditional SEO, digital media, social, and content with cutting-edge generative AI capabilities, including its deep learning engine, DataMind, and AI Catalyst platform. Trusted by enterprises, mid-market companies, and leading digital agencies, BrightEdge continues to set the standard for innovation in search and AI, enabling brands to win by becoming an integral part of the digital experience. 

Contact: press@brightedge.com 

Press Release Date

How Google AI Overviews and ChatGPT Cite Wikipedia Differently

Both engines cite the world's most-referenced source. But the company Wikipedia keeps, and when it gets left out entirely, reveals how differently each engine defines authority.

Both Google AI Overviews and ChatGPT cite Wikipedia across a wide range of queries. Both engines clearly trust it. But trust is only part of the story. When you look at what Wikipedia is cited alongside in each engine, and the queries where it ranks first organically but still doesn't appear in the AI layer, a more nuanced picture emerges about how each platform actually defines authority and when it chooses to use it.

We used BrightEdge AI Hypercube and DataCubeX to analyze the prompts where Wikipedia is cited across both Google AI Overviews and ChatGPT, the sources that appear alongside it in each engine's responses, and the organic ranking data for tens of thousands of keywords where Wikipedia ranks and an AI Overview is present. The patterns are consistent and instructive.

Data Collected

 

Data PointDescription
Co-citation analysisFor every prompt where Wikipedia was cited, all other brands and domains cited in the same response were extracted and categorized by source type across both platforms
Co-occurrence ratesCalculated as the percentage of Wikipedia-cited prompts that also included each named source in the same response
Organic rank vs. AIO citationCross-referenced Wikipedia's organic ranking position against whether it appeared as a cited source in the AIO for the same keyword, across tens of thousands of keywords
Citation rate by rank tierAIO citation rates segmented by Wikipedia's organic ranking position (top 3, 4-5, 6-10, 11-20, 21+)
Exclusion pattern analysisKeywords where Wikipedia holds a top-3 organic position but does not appear in the AIO, analyzed for query type patterns

 

Key Finding

Google AI Overviews and ChatGPT cite Wikipedia in fundamentally different contexts. In AIO, Wikipedia sits alongside social platforms and community sources. In ChatGPT, it sits alongside institutional authorities and credentialed reference sources. Same citation. Two completely different signals about what each engine thinks authoritative means. And separately, even holding the #1 organic position isn't sufficient for AIO inclusion. The format of the content has to match what the query actually needs.

 

Same Source. Completely Different Neighborhoods.

When Wikipedia is cited in a response, it doesn't appear alone. Both engines surface multiple sources per response, and the pattern of what appears alongside Wikipedia is strikingly different between the two platforms.

In Google AI Overviews, the most common sources cited in the same response as Wikipedia are YouTube, Reddit, and Quora, appearing in 13%, 9%, and 6% of Wikipedia-cited AIO responses respectively. The broader co-citation set includes news outlets, entertainment indexes, sports media, and community discussion platforms. Wikipedia in this context is functioning as a credibility anchor in a broad, socially-validated ecosystem.

In ChatGPT, the landscape is almost entirely different. Encyclopedia Britannica appears in the same response as Wikipedia in 43% of ChatGPT responses. Merriam-Webster appears in 13%. The remainder of the top co-citations are health publishers, legal reference institutions, and scientific databases.

AIO puts Wikipedia in the company of platforms where people engage with content. ChatGPT puts it in the company of sources people use to verify it. Same citation. Two completely different competitive sets.

Top Co-Citations Alongside Wikipedia - Google AI Overviews

SourceCo-occurrence RateType
YouTube13%Video platform
Reddit9%Community discussion
Britannica7.5%Reference encyclopedia
Quora6%Q&A community
IMDb5.2%Entertainment index
Facebook3.8%Social platform

Top Co-Citations Alongside Wikipedia - ChatGPT

SourceCo-occurrence RateType
Encyclopedia Britannica43%Reference encyclopedia
Merriam-Webster13%Dictionary / reference
Cleveland Clinic6.3%Health institution
Healthline5.4%Health publisher
Mayo Clinic4.7%Health institution
Reddit3.3%Community discussion

 

Authority and Citation Are Two Different Decisions.

Across tens of thousands of keywords where Wikipedia holds an organic ranking and an AI Overview is present, Wikipedia makes it into the AIO on fewer than half of those queries. That gap is worth examining closely, because the exclusions aren't random.

When Wikipedia is cited in AIO, 75% of the time it holds a top-3 organic ranking. Median organic position: 2. When Wikipedia is not cited, roughly a third of those cases still have Wikipedia sitting at position #1 organically.

The exclusion pattern reveals why. For live sports queries, real-time events, and navigational searches, Wikipedia holds pages on those topics, but AIO needs a live data feed, not a reference article. The content format doesn't fit the query's immediate need, regardless of how authoritative the domain is. A similar logic applies to certain sensitivity-adjacent topics and queries with strong navigational intent.

Ranking reflects topical authority. AIO citation reflects whether the content format can directly serve what the query needs right now. For Google, those are two separate decisions.

 

What Marketers Need to Know

The competitive set depends on which engine you're in. If you're competing on queries where Wikipedia appears in AIO, you're competing alongside social platforms, community content, and entertainment sources. If you're competing on queries where Wikipedia appears in ChatGPT, you're competing alongside institutional reference authorities. These require different content investments.

Ranking reflects authority. AIO citation reflects usefulness. Google can rank Wikipedia #1 and still not include it in the AIO, because ranking rewards topical credibility while AIO asks a different question: can this content directly answer what the user needs right now? For real-time and navigational queries, an encyclopedia entry can't, regardless of how authoritative the domain is.

The co-citation data tells you who else is in the room. For any query set where Wikipedia shows up, the co-citation patterns give you an accurate picture of the competitive landscape inside that AI response. That competitive set looks fundamentally different in AIO versus ChatGPT, and mapping it for your own category is the starting point for a differentiated citation strategy across both engines.

Content format is a citation variable. AI Overviews make active judgments about whether a piece of content is the right format to answer a specific query type. Authoritative content that isn't structured to serve the query's immediate need may rank highly and still be excluded from the AI layer.

 

Technical Methodology

ParameterDetail
Data SourcesBrightEdge AI Hypercube (prompt-level co-citation analysis); BrightEdge DataCubeX (organic ranking vs. AIO citation cross-reference)
Engines AnalyzedGoogle AI Overviews, ChatGPT
Query SetTens of thousands of prompts where Wikipedia appears as a cited source across both platforms; separately, tens of thousands of keywords where Wikipedia holds an organic ranking and an AI Overview is present
Co-occurrence CalculationNumber of Wikipedia-cited responses also citing that source divided by total Wikipedia-cited responses
Citation Rate AnalysisAIO citation defined as Wikipedia's URL appearing as a source in the AI Overview. Non-citation defined as AI Overview present, Wikipedia ranking organically, URL not appearing in AIO sources.

 

Key Takeaways

FindingDetail
AIO's Wikipedia neighborhood is socialYouTube (13%), Reddit (9%), and Quora (6%) are the most common co-citations. The ecosystem skews toward community engagement and social content.
ChatGPT's Wikipedia neighborhood is institutionalEncyclopedia Britannica appears in 43% of ChatGPT responses that also cite Wikipedia. Merriam-Webster at 13%. The ecosystem skews toward credentialed reference authorities.
Organic rank predicts but doesn't guarantee AIO citation75% of Wikipedia's AIO citations come from a top-3 organic ranking. But roughly a third of exclusion cases still have Wikipedia at position #1.
Content format determines AIO inclusionReal-time, navigational, and sensitivity-adjacent queries produce consistent exclusion patterns regardless of organic authority.
Two engines, two competitive setsThe co-citation data maps who brands are competing against for AI visibility, and that map looks entirely different in AIO versus ChatGPT.

Download the Full Report

Download the full AI Search Report — How Google AI Overviews and ChatGPT Cite Wikipedia Differently

Click the button above to download the full report in PDF format.

Published on  April 2, 2026

APAC Webinar: Optimizing for AI Agents – What Marketers Need to Know About Crawl Behavior

Ensure your content is discoverable, usable, and preferred in AI-powered search experiences

This session goes live on Wednesday, April 22, 2026, at 1 PM AEDT . Reserve your spot and get ahead of how AI agents are transforming the way your content is found and used.

Description

AI agents are no longer a future concept. They are actively crawling your website right now. From ChatGPT's operator agents to Perplexity, Google's Gemini agents, and beyond, a new class of automated visitors is determining how your brand shows up in agentic search experiences. And they don't behave like traditional search bots. But is your site technically prepared for how these agents access, parse, and act on your content? 

Join us for this technical deep dive into AI agent crawl behavior and what it means for your marketing strategy. While foundational SEO practices still apply, the agentic era introduces a new layer of considerations — and knowing how to communicate their business impact internally is critical for securing the resources to act on them. We'll walk you through what you need to know to ensure your content is accessible, actionable, and preferred by AI agents across the search landscape.

What You'll Learn

  • How AI agent crawlers differ from traditional search bots — and why it matters
  • The technical factors that influence whether agents can access and use your content
  • How to audit your site for AI agent readiness
  • How to make the internal case for prioritizing agent optimization

Featured Speakers:

Kylie Tabrett
 
 
Having trouble with the form? Please try refreshing the page or contact us directly at support@brightedge.com

Save your spot 

* indicates required
 
 
 

 
 
 
 
×
Table of Contents
Table of contents

AI Agents for Search Marketers

Your website has a new set of visitors, and most marketing teams have no idea they are there. AI agents from OpenAI, Anthropic, Google, Apple, Perplexity, and others are crawling the web at scale, reading your content, and deciding in real time whether to cite it, summarize it, or skip it entirely. According to BrightEdge research, AI agent crawl activity across enterprise sites spans three distinct categories: agents crawling for training data, agents powering real-time search answers, and agents acting on behalf of individual users. Together they represent a new layer of traffic that are invisible for traditional analytics, responds to different signals than Googlebot, and increasingly determines where your brand appears in AI-generated answers.

This guide covers everything search marketers and digital teams need to understand about AI agents: who they are, what they crawl, how to make sure they can reach your content, and how to optimize for visibility in the AI-powered search landscape that is already here.

 

What Are AI Agents and Why Are They Different?

The term "AI agent" gets used loosely, which creates confusion for marketers trying to understand what is actually happening on their sites. For the purposes of search and content visibility, an AI agent is a crawler that navigates the web, retrieves and processes content, and uses what it finds to power AI technologies such as chatbots, search engines, research tools, or to train Large Language Models. The agent is not a user. It does not convert, it does not bounce, and it never shows up in your GA4 sessions report.

That last point is critical. Most of the traffic that AI agents generate is invisible in standard analytics because agents do not execute JavaScript, do not trigger client-side tracking tags, and do not carry session cookies. The most reliable way to see them is through server logs or a purpose-built intelligence layer such as BrightEdge's AI Agent Analytics.

How AI Agents Differ From Traditional Web Crawlers

Googlebot has been crawling the web since 1998. Marketing teams have spent decades learning to accommodate it: sitemaps, robots.txt, canonical tags, structured data. AI agents operate on a fundamentally different model. Understanding the differences is the starting point for optimization.

Purpose

Googlebot crawls to index and rank. AI agents crawl to understand, synthesize, and cite. But before that distinction matters, something more fundamental has to be true: the AI agent has to know your site exists, and it has to decide your pages are worth reading.

That determination starts with traditional search. Every major AI agent (GPTBot, ClaudeBot, Gemini, Perplexity) uses a traditional search index as part of its sourcing infrastructure. If your pages aren't indexed, aren't earning authority, and aren't signaling topical relevance through the fundamentals, you are invisible to AI agents before the question of citation ever comes up.

Traditional search visibility is the price of admission to AI search.

Once you're in the room, the signals diverge. A page that ranks in the top ten organic results will not automatically earn an AI citation, and a page that is frequently cited in AI-generated answers may not appear in the top ten at all. The overlap is real but incomplete, and understanding where it breaks down is where the optimization opportunity lives.

Content Consumption

Traditional crawlers primarily care about structure: title tags, headers, internal links, page speed. AI agents are trained on high-quality human-curated text, which suggests they respond to the same signals that make content worth reading: clear arguments, specific claims, credible sourcing. The exact weighting of those signals in citation selection is not publicly documented by any platform. What is observable is that content demonstrating genuine expertise on a topic consistently performs better in citation than content covering the same subject superficially.

JavaScript and Rendering

Most AI agents do not render JavaScript. If your key content, navigation, or product details live behind client-side rendering, a meaningful portion of AI agents simply cannot see it. This is not a theoretical risk. It is a structural gap that affects a large share of modern websites, particularly those running heavy frontend frameworks without server-side rendering.

Crawl Behavior

Googlebot crawls with a relatively predictable schedule tied to PageRank and crawl budget. AI agents crawl far less predictably. They respond to user demand: when a topic spikes in AI chat activity, the relevant agents often re-crawl the web to update their knowledge. According to BrightEdge research, AI training crawl activity grew by more than 160% in a single month in late 2025, driven by large-scale re-indexing events. This kind of volatility has no equivalent in traditional SEO.

What AI agents do share with traditional crawlers is a reliance on robots.txt and XML sitemaps. Despite all the discussion around llms.txt as an emerging standard for AI-specific directives, most AI agents currently do not request it. What they do request is robots.txt, which means your existing access rules are already governing whether these agents can reach your content. They also pull sitemaps, making an up-to-date sitemap one of the simplest and most overlooked levers for AI crawl coverage. If your sitemap is stale or incomplete, you are leaving pages off the table before an agent ever decides whether to cite them.

Three Types of AI Agents

Not all AI agents serve the same function. BrightEdge research identifies three distinct categories of AI agent crawl activity, each with different implications for search marketers.

  • Training agents build and refresh the knowledge base that powers AI models. They crawl broadly and at high volume. Their activity does not directly drive citations, but the content they consume shapes what the model knows and, over time, what it surfaces in answers. According to BrightEdge research, training crawls account for the largest single share of AI agent activity on enterprise sites.
  • Search agents retrieve content to power AI search answers and citations. When someone asks Perplexity a question or uses ChatGPT Search, these agents go out and fetch current information. They are citation-driven by design, and optimizing for them has the most direct impact on AI search visibility.
  • User agents act on behalf of an individual user during an active session — powering browse mode in ChatGPT, research features in Claude, or AI-assisted tools like Google NotebookLM. They crawl fewer pages per session but often with high intent, following links in ways that reflect a real research workflow.


     

Which AI Agents Are Crawling Your Site?

The AI agent ecosystem has expanded faster than most marketing teams have tracked it. What started as a handful of well-known bots has grown into a multi-platform landscape spanning major technology companies, independent AI search engines, and emerging players whose crawl volume is rising quickly.

The table below covers the agents BrightEdge research tracks across sites, including their user-agent strings (the identifiers you will find in your server logs and robots.txt configurations), the products they feed, and their relative share of AI crawl activity.

Agent NameUser-Agent StringTypePowersBrightEdge Data
GPTBotopenai_gptbotTrainingChatGPT, Responses API51% of training crawls
OAI-SearchBotOAI-SearchBotSearchChatGPT Search49% of AI search crawls
ChatGPT Useropenai_chat_gpt_userUserChatGPT (browse mode)96% of AI user traffic
ClaudeBotClaudeBotTrainingClaude.ai, Claude API17% of training crawls
Claude SearchBotClaude-SearchBotSearchClaude (web search)15% of AI search crawls
PerplexityBotPerplexityBot/1.0SearchPerplexity AI answers6% of AI search crawls
AppleBotApplebotSearchApple Intelligence, Siri29% of AI search crawls
ByteSpiderBytespiderTrainingTikTok, ByteDance AIUp 138% Nov–Feb 2026
Google-ExtendedGoogle-ExtendedTrainingGemini, AI OverviewsSeparate from Googlebot
NotebookLM BotGoogle-NotebookLMUserGoogle NotebookLMUp 144% Nov–Feb 2026
DuckAssistBotDuckAssistBotUserDuckDuckGo AI ChatConsistent activity
MetaExternalAgentMeta-ExternalAgentTrainingMeta AILow volume, emerging

Source: BrightEdge AI Agent Insights research, November 2025 through February 2026.

What the Data Tells Us

OpenAI Has the Largest Footprint — Across All Three Categories

According to BrightEdge research, OpenAI agents account for the majority of AI agent crawl activity across enterprise sites when training, search, and user categories are measured together. GPTBot (the training crawler) represents more than half of all AI training crawl volume. OAI-SearchBot drives nearly half of all AI search crawl activity. And ChatGPT's user-facing agent accounts for more than 96% of all AI user bot traffic. No other platform comes close to that concentration across all three functions simultaneously.

For search marketers, this means that if your site has any friction with OpenAI agents — whether through robots.txt blocking, server configuration issues, or JavaScript rendering barriers — the downstream impact on ChatGPT Search visibility is significant.

Apple Is a Bigger Player Than Most Teams Realize

Applebot accounts for nearly 30% of AI search crawl activity in BrightEdge research, making it the second-largest AI search crawler by volume. Apple Intelligence, Siri's enhanced web features, and Safari's AI summaries all run through this agent. Because Apple does not have a traditional search engine interface, many teams overlook its role in AI-driven content distribution. That is a blind spot worth correcting.

ByteSpider Is Growing Fast

Bytespider, the crawler powering TikTok's parent company ByteDance and its AI products, grew its crawl volume by 138% between November 2025 and February 2026 in BrightEdge research. It now represents nearly a third of all AI training crawl activity. Most marketing teams are not tracking it at all. As TikTok's AI search and answer features expand, ByteSpider's role in determining content eligibility will grow with them.

ClaudeBot Spiked Dramatically in Late 2025

Anthropic's training crawler (ClaudeBot) saw an increase of more than 800% between November and December 2025, followed by continued elevated activity through early 2026. This kind of spike typically reflects a large-scale model training or knowledge refresh event. It is a reminder that AI training crawl activity does not follow a linear schedule — it surges in response to platform decisions that happen entirely outside your control. Visibility into that activity requires real-time monitoring, not after-the-fact log analysis.

Google NotebookLM Is an Emerging Signal

Google NotebookLM's crawl agent grew by 144% between November 2025 and February 2026 in BrightEdge research. While its total volume remains smaller than the major search and training crawlers, its trajectory reflects the broader expansion of AI user tools that fetch live web content on behalf of researchers, analysts, and knowledge workers. Content that is accessible and well-structured has a direct advantage as this category scales.

 

 

A Note on Agent Identification

User-agent strings are declared by the crawler but are not cryptographically verified. Most major agents follow declared conventions, but you should validate agent identity through their documented IP ranges when it matters. OpenAI, Anthropic, Google, and Perplexity all publish their IP ranges and crawler documentation. Cross-referencing user-agent strings with source IP ranges is the reliable method for confirming agent identity in your server logs.

Are You Blocking the Agents That Matter?

A meaningful share of enterprise sites are inadvertently blocking one or more major AI agents. They are not doing it on purpose. They are doing it through legacy robots.txt configurations, blanket AI-blocking rules added reactively in 2023 and 2024, or misconfigured server settings that treat AI agents the same way they treat scrapers. The result is the same regardless of intent: invisible to AI search, ineligible for citation, absent from AI-generated answers.

According to BrightEdge research, server errors accounted for nearly a quarter of all ChatGPT user agent requests in January 2026 across a broad set of enterprise sites. That suggests configuration problems rather than isolated incidents.

Understanding robots.txt for AI Agents

The robots.txt file remains the primary mechanism for controlling how non-human entities access your site. The challenge is that most robots.txt files were written for a world with two or three major crawlers, not the dozen or more AI agents now active on the web. Rules that made sense in 2020 may be creating unintended blocks today.

The Wildcard Problem

The most common configuration issue is a wildcard Disallow rule used to block the generic "everyone else" catch-all, combined with explicit Allow rules for Googlebot and a few other crawlers. In theory, this lets the crawlers you want in while blocking the ones you do not. In practice, if you have not explicitly listed every major AI agent, the wildcard catches them all.

# Common configuration that inadvertently blocks all AI agents
User-agent: Googlebot
Allow: /

User-agent: Bingbot
Allow: /

User-agent: *
Disallow: /

# GPTBot, OAI-SearchBot, ClaudeBot, Applebot, PerplexityBot
# and every other AI agent falls through to the wildcard Disallow

If your robots.txt looks similar to the example above, every AI agent not explicitly named is blocked. That includes ChatGPT Search, Claude, Perplexity, and Apple Intelligence.

The Reactive Block Problem

In 2023 and 2024, many brands added explicit blocks for AI training crawlers in response to legitimate concerns about content scraping and unauthorized use of proprietary material. The problem is that many teams blocked agents using identifiers that also cover the search and user-facing versions of those same products. Blocking GPTBot blocks OpenAI training crawls. Blocking OAI-SearchBot blocks ChatGPT Search. They are different agents with different functions, governed by different user-agent strings, and the distinction matters more than most teams realize.

This is critical because those agents power the AI search that connects customers to products and services they need. This includes users that research software vendors, compare service providers, evaluate products before purchase, and identify partners for enterprise deals. These are purchase-intent moments that used to drive a user directly to your site from organic search. If your content is blocked from the agents powering those interactions, you risk not being in that conversation — and a competitor whose content is accessible will be.

The Recommended Configuration

Most enterprise sites should keep AI agents open by default. Restricting access is likely costing you visibility.

# Recommended baseline for AI agent access

# OpenAI: allow search and user agents, manage training separately
User-agent: OAI-SearchBot
Allow: /

User-agent: openai_chat_gpt_user
Allow: /

User-agent: GPTBot
Allow: / # Change to Disallow: / if you want to opt out of training

# Anthropic
User-agent: ClaudeBot
Allow: /

User-agent: Claude-SearchBot
Allow: /

# Apple
User-agent: Applebot
Allow: /

# Perplexity
User-agent: PerplexityBot
Allow: /

# Google AI (separate from Googlebot)
User-agent: Google-Extended
Allow: / # Change to Disallow: / to opt out of Gemini/AI Overviews training

Beyond robots.txt: Other Access Barriers

Server-Level Blocks

Some hosting configurations, WAF (Web Application Firewall) rules, and bot management platforms treat AI agents the same as malicious scrapers. If your site has Cloudflare, Akamai, Imperva, or a similar bot management layer, check whether AI agents are being challenged or blocked at the network level. A 200-status response in your robots.txt is irrelevant if the agent is getting a 403 or a CAPTCHA challenge before it reaches your content.

The January 2026 server error spike in BrightEdge research — where nearly one in four ChatGPT user agent requests returned a server error — is consistent with WAF or rate-limiting rules that tightened during a period of elevated AI crawl activity. Sites that addressed this proactively maintained their citation visibility. Sites that did not saw gaps.

Brightedge AI Agent Insights showing only 70% of agent visits having successful responses, with 5 specifically encountering server side errors

JavaScript Rendering

Most AI agents do not execute JavaScript. This means content loaded client-side — product descriptions, article bodies, navigation, or metadata — may simply not exist from an AI agent's perspective. If your site uses React, Next.js, Angular, or a similar framework without server-side rendering or static generation, run a crawl simulation using a non-JavaScript crawler and compare what it sees to what a browser renders. The gaps are where AI agents are going blind.

Crawl Rate and Response Speed

AI agents do not always send a Crawl-Delay directive or respect one if present. Sites with aggressive rate limiting may be returning errors to AI agents that are crawling at a pace the server interprets as suspicious. Review your rate-limiting thresholds and confirm that known AI agent IP ranges are not being throttled or blocked by your infrastructure. Each major AI platform publishes its crawler IP ranges; adding them to an allowlist in your WAF or CDN is a straightforward step.

How to Audit Your Current AI Agent Access

If you are not sure whether your site has AI agent access issues, here is a practical audit process to run today.

  1. Check your robots.txt. Fetch yourdomain.com/robots.txt and look for any Disallow: / rules, wildcard configurations, or explicit blocks for AI agent user-agent strings. Flag anything that could be catching agents you want to allow.
  2. Audit crawl activity. Pull the last 30 days of AI agent activity and review which agents are reaching your site, their request volume by category, and their error rates by agent type. A healthy AI agent interaction should show 200-status responses at 90% or better.
  3. Check your WAF and CDN rules. Review bot management configurations for any rules that apply to AI agent user-agent strings or IP ranges. Confirm that known AI crawler IP ranges are not on block or challenge lists.
  4. Validate your sitemap coverage. AI search agents use sitemaps as a starting point. Confirm your XML sitemap is current, correctly submitted, and includes the pages you most want cited. Exclude pages you do not want agents to reach.

llms.txt Is an Emerging Format but Slow to Adopt

llms.txt is an emerging specification that lets site owners create a curated index of their most important content for AI agents. The idea is sound: rather than forcing an agent to crawl and parse HTML, you give it a clean, structured entry point to your most important pages.

Adoption among publishers has grown quickly as Anthropic, Cloudflare, Stripe, Perplexity, and others have all published them. The catch is that there is currently no evidence the major AI agents are systematically requesting llms.txt files. Publishing one is low effort and worth doing as a forward-looking measure, but it won't move any needle for you today.

Monitor this space. If agent behavior shifts and retrieval patterns start showing llms.txt requests at scale, the implementation lift is minimal and you will want to be ready.

What llms.txt Is and What It Is Not

According to the official specification maintained at llmstxt.org, the file is designed primarily for use at inference time: when a user is actively asking an AI tool for help and the agent is fetching context to answer well. It is a table of contents, not a content delivery mechanism. The linked pages still need to serve accessible, well-structured content for the agent to actually read.

The llms.txt File Format

The file lives at the root of your domain at yourdomain.com/llms.txt. It uses a specific markdown structure defined by the spec:

  • H1 header: The name of your site or project. This is the only required element.
  • Blockquote: A short summary of the site or project. One to two sentences capturing what the site is and who it serves.
  • Optional body text: Additional context about the site, its content, or how to interpret the files below. No heading elements here.
  • H2 sections with file lists: Named sections, each containing a markdown list of links with optional descriptions. These point to the most important pages on your site.
  • Optional section: A special H2 section labeled "Optional" that signals to agents they can skip those links if context window space is limited. Use it for secondary content.
# Your Company Name
> Brief summary of what your company does and who it serves.

Additional context about this site and how to interpret the content below.

## Products
- [Product Name](https://yoursite.com/products/name.md): What this product does and who it is for
- [Product Documentation](https://yoursite.com/docs/product.md): Full technical documentation

## Resources
- [Research Reports](https://yoursite.com/research/index.md): Original data and industry research
- [Case Studies](https://yoursite.com/case-studies.md): Client outcomes and implementation examples

## Optional
- [About Us](https://yoursite.com/about.md): Company background and leadership
- [Blog](https://yoursite.com/blog/index.md): Ongoing thought leadership content

A companion file, llms-full.txt, embeds the full content of all linked pages directly into a single file rather than requiring an agent to follow each link. This is particularly useful for developer tools, AI coding assistants, and agents that need comprehensive context in a single request.

Serving AI-Friendly Markdown Alongside HTML

The llms.txt spec also proposes that individual pages serve a clean markdown version at the same URL with .md appended. So yoursite.com/about would also be accessible at yoursite.com/about.md. This gives agents a direct path to clean, token-efficient content without having to parse navigation menus, footers, CSS classes, and JavaScript that carry no semantic value for an AI reader.

The efficiency difference matters. Cloudflare, which has built native markdown serving support into its platform, notes that a simple heading in markdown costs roughly three tokens, while its HTML equivalent with class attributes, div wrappers, and script tags can cost four to five times as many. Across a full page, that gap compounds significantly. Agents with limited context windows make coverage tradeoffs, and pages that waste tokens on structural HTML get less of their actual content read.

Cloudflare's "Markdown for Agents" feature implements this through HTTP content negotiation: when an agent sends an Accept: text/markdown header, the server returns clean markdown. Browsers receive normal HTML from the same URL. This is not cloaking — it is serving the same content in different formats to different consumers, which is a standard web practice.

Content Formatting for AI Agents

Once you have confirmed that AI agents can reach your content, the next question is whether that content is formatted in a way agents can efficiently read, understand, and cite.

HTML and AI Agents

Standard HTML pages carry a significant amount of structural content that is essential for browser rendering but carries no semantic value for an AI agent. Navigation menus, footer links, CSS class names, JavaScript function calls, schema markup wrapper tags, and advertising containers all consume tokens without contributing to the meaning of your content. An agent reading a typical enterprise webpage spends a large portion of its context window on markup that tells it nothing about what you know or what you offer.

The practical result is that agents get less of your actual content per request. Pages that efficiently deliver their meaning in clean, structured text have an advantage in how thoroughly they get read and processed — which is the precondition for citation.

Writing for AI Agent Comprehension

The formatting practices that make content clear to AI agents align closely with what makes content clear to human readers. That is not a coincidence. Agents learn from human-written text and respond to the same clarity signals humans do. The following principles are drawn from how AI agents are known to process and evaluate text.

Lead With the Answer

AI agents, like human researchers, are often looking for a direct answer to a specific question. Content that buries the key claim or conclusion three paragraphs in makes the agent work harder to extract it — and risks the most important information being truncated when context windows fill up. Put the primary point in the first sentence of each section, then support it.

Use Headers to Signal Structure

Markdown H1, H2, and H3 headers are strong semantic signals. An agent reading your content uses headers to build a structural map of the page before processing the body text. Clear, descriptive headers that accurately summarize the section that follows help agents navigate and excerpt content appropriately. Vague headers ("More Information," "Overview," "Details") reduce that navigational value.

Write in Discrete, Self-Contained Units

Agents often read and cite individual passages rather than full pages. A paragraph or section that can stand alone, without requiring the reader to have read the previous three paragraphs for context, is easier to cite accurately. This does not mean writing in fragments — it means that each section should be coherent on its own as well as part of the whole.

Use Lists for Enumerable Content

When content consists of discrete items — steps in a process, features of a product, categories in a taxonomy — bullet or numbered lists communicate that structure more clearly than prose. An agent reading a list knows it is looking at a collection of parallel items. The same information in paragraph form requires the agent to infer that structure, which introduces parsing overhead and potential for misrepresentation.

Be Consistent With Terminology

Agents build an understanding of your domain vocabulary from the text they read. If your site refers to the same concept by three different names across different pages — using "customer," "client," and "user" interchangeably, for example — agents may treat them as distinct concepts or fail to connect related information. Standardizing terminology across your content library is a low-effort, high-impact practice for AI comprehension.

Avoid Jargon Without Definition

Industry shorthand that your human audience recognizes instantly may mean nothing to an agent operating outside your specific context. If you use acronyms or proprietary terminology, define them on first use and use them consistently thereafter. This applies equally to product names, internal frameworks, and category labels.

Schema Markup and Structured Data

JSON-LD structured data remains relevant for AI agents, though its role differs from its function in traditional SEO. Search engines use schema markup primarily for rich result eligibility. AI agents use it to confirm entity identity: who wrote this, what type of content it is, when it was published, what organization produced it, and how it relates to other entities on the web.

The most impactful schema types for AI agent comprehension are:

  • Article / NewsArticle / BlogPosting: Confirms content type, authorship, and publication date. Date signals matter — agents that prefer fresh content can use dateModified to assess recency.
  • Organization: Establishes your entity identity, official name, website, and related properties. This is the schema that helps agents correctly attribute content to your brand across different pages and platforms.
  • Person: Author credentialing. Agents increasingly weight the demonstrated expertise of the person behind content. Linking author markup to a consistent entity across your site strengthens that signal.
  • FAQPage and HowTo: These types present content in a format that maps directly to how AI search agents construct answers. Well-marked FAQ and HowTo content is structurally pre-formatted for citation.
  • Product: For ecommerce and product pages, structured product data including name, description, price, and reviews helps agents present accurate product information in AI shopping answers.

AI Agent Optimization Checklist

Access and Crawl Configuration

  • robots.txt reviewed — no wildcard Disallow rules blocking AI agents without explicit Allow exceptions
  • All major AI search agents explicitly listed in robots.txt: OAI-SearchBot, Claude-SearchBot, Applebot, PerplexityBot
  • Training agent policy documented and intentional: GPTBot and Google-Extended allowed or blocked based on stated content policy
  • WAF and CDN bot management rules reviewed — AI agent IP ranges on allowlist, not subject to generic bot-blocking
  • Rate limiting thresholds reviewed — known AI crawler IP ranges not being throttled
  • Server logs or AI Agent Insights reviewed to confirm major agents are reaching site with 90%+ 200-status response rate

Structured Data

  • JSON-LD structured data present on all article and blog pages (Article or BlogPosting schema)
  • Organization schema present on homepage with official name, URL, and sameAs links
  • Author schema present on key content pages with consistent Person entity
  • FAQPage schema implemented where FAQ content exists
  • Product schema complete and accurate for ecommerce pages

Frequently Asked Questions

What is an AI agent in the context of SEO?

An AI agent is an automated system that crawls the web, retrieves and processes content, and uses what it finds to power an AI product: a chatbot, a search answer engine, a research tool, or a training dataset. Unlike a traditional web crawler focused on indexing pages for ranking, AI agents are evaluating content for use in answers, citations, and model training.

How is an AI agent different from a chatbot?

A chatbot is the user-facing product. An AI agent is the system that goes out and gets information to power that product. When you ask ChatGPT a question with web search enabled, ChatGPT is the chatbot — OAI-SearchBot is the agent that goes and retrieves current web content to inform the answer. The terms are often used interchangeably, but for SEO purposes the distinction matters: the agent is what is crawling your site.

Does robots.txt work for AI agents?

Most major AI agents follow robots.txt conventions. OpenAI, Anthropic, Google, Perplexity, and Apple have all documented their crawlers and stated they respect robots.txt. However, compliance is voluntary. You cannot force any agent to adhere to robots.txt, though you can verify agent behavior through server logs and follow up with the platform if violations occur.

Should I block AI agents to protect my content?

This depends on your goals and the specific agent. Blocking AI training crawlers (GPTBot, Google-Extended, ClaudeBot) prevents your content from being used in model training and is a legitimate choice. Blocking AI search and user agents (OAI-SearchBot, Claude-SearchBot, Applebot, PerplexityBot) removes your content from eligibility for AI search citations and AI-generated answers, which is increasingly a significant traffic and visibility channel. Most brands benefit from a differentiated policy: block or restrict training agents based on content policy, while permitting search and user agents.

What is GPTBot and should I allow it?

GPTBot is OpenAI's training crawler. It crawls the web to collect content for model training. Allowing GPTBot means your content may be used to train future OpenAI models. Blocking it means it will not be, but it does not affect whether ChatGPT Search (powered by OAI-SearchBot) can cite you in live answers. These are separate agents with separate robots.txt user-agent strings. Check yourdomain.com/robots.txt and confirm you have an explicit policy for both.

What is the difference between GPTBot and OAI-SearchBot?

GPTBot is OpenAI's training crawler — it builds the underlying knowledge of the model. OAI-SearchBot is OpenAI's real-time search crawler — it retrieves current web content to answer live user queries in ChatGPT Search. Blocking GPTBot affects training data. Blocking OAI-SearchBot affects live search citation. Most visibility-focused brands want to allow OAI-SearchBot even if they restrict GPTBot.

What is llms.txt and do I need one?

llms.txt is a markdown file proposed in September 2024. It lives at yourdomain.com/llms.txt and provides AI agents with a curated map of your most important content. It is most useful for sites with developer documentation, APIs, or structured product information. For content-first marketing sites, it is worth implementing as a good-practice signal, but access and content quality are higher priorities.

Why is my AI agent traffic not showing up in Google Analytics?

Most AI agents do not execute JavaScript, which means they do not trigger GA4 or any other client-side analytics. Server logs capture agent activity because they record all HTTP requests regardless of JavaScript execution. AI referral traffic from users who click through from AI answers does appear in analytics — look for referral sessions from domains like perplexity.ai, chatgpt.com, and similar AI platform domains.

How do I know if AI agents are encountering errors on my site?

Check your server logs for 4xx and 5xx status codes associated with AI agent user-agent strings. A healthy AI agent interaction should show 200-status responses on 90% or more of requests.

Does having a faster website help with AI agent crawling?

Yes, particularly for search-type agents that are operating under time constraints when fetching content for live user queries. Slow server response times can result in timeouts or partial content retrieval.

How do I get my content cited in AI answers?

The observable factors that correlate with stronger citation performance are: clean agent access, well-structured content, consistent entity signals, depth of expertise on your core topics, and regular content updates.

How is AI search different from traditional SEO?

Traditional SEO optimizes for ranking positions in a list of results. AI search optimization aims for citation presence in AI-generated answers that may not include a traditional ranked list at all. Some signals overlap — such as authority, quality, and relevance — but the specific content formatting, technical requirements, and measurement approaches are distinct. This guide includes specific technical components that are unique to Agentic Engine Optimization (AEO).

×
Request a demo

Request a Personalized Demo

* indicates required
 
 
 

How Google AI Overviews and ChatGPT Cite Retailers Differently

When someone's ready to buy, these two platforms take very different paths to an answer.

When someone's ready to buy, these two platforms take very different paths to an answer.

AIO operates inside a commerce-ready SERP. ChatGPT is the whole page. That architectural difference shapes everything about how each platform handles purchase-intent queries and which brands get cited.

When a consumer types a purchase-intent query into Google or ChatGPT, both platforms are trying to do the same thing: give a useful answer. But the path each takes looks remarkably different. And the difference isn't really about the AI. It's about what's around it.

Google AI Overviews sit on top of a SERP that already has Shopping carousels, merchant listings, and organic results. The AI doesn't need to close the transaction on its own. ChatGPT is the whole page. No carousel. No product listing unit. No organic fallback. When someone asks ChatGPT something with purchase intent, the AI has to do all the work itself, including the evaluative work that a full SERP would otherwise distribute across multiple surfaces.

We used BrightEdge AI Hyper Cube to analyse tens of thousands of prompts where the top U.S. retailers appear, tracking mentions, citations, and brand sentiment across both Google AI Overviews and ChatGPT. We filtered to transactional intent to understand how each platform behaves when someone is ready to buy.

The behaviour gap is real. And the architecture explains almost all of it.

Data Collected

 

Data PointDescription
Citation volume by platformTotal query count where major retailer domains were cited as sources in Google AI Overviews vs. ChatGPT
Transactional intent filteringPrompts filtered and cross-referenced by purchase intent across both platforms
Citation source classificationEach cited domain categorized by type: major retailer, social/community, editorial/financial, news media, government/academic, other/niche
Brand mention trackingAll brand mentions extracted from AI responses and classified by sentiment: positive, neutral, negative
Competitive set analysisAverage number of brands surfaced per transactional response on each platform
Cross-platform comparisonHead-to-head citation intent and source analysis across both engines using matched query methodologies

 

Key Finding

Google AI Overviews and ChatGPT handle retail and purchase-intent queries in fundamentally different ways. Not because they have different goals, but because the environments they operate in are fundamentally different. AIO can lean on the SERP's existing commerce infrastructure to do the transactional heavy lifting. ChatGPT cannot. That single architectural distinction drives measurable differences in which sources get cited, how many brands get surfaced, and how often negative sentiment appears in the response.

 

Start With the Environment, Not Just the AI

Understanding the behavioral differences between these two platforms starts with understanding what each platform is embedded in.

Google AI Overviews appear within a search results page that already contains Shopping carousels, merchant product listings, local results, and organic links. A user who sees an AIO response has immediate access to purchase options below it. The AI can gesture toward a retailer, cite their domain, reference their pricing, surface their brand, and the SERP infrastructure does the rest.

ChatGPT has none of that. The response is the experience. If a user wants to act on a ChatGPT recommendation, the AI needs to provide enough evaluative context to justify the action. There's no carousel to fall back on. No organic listing to validate the pick. The AI is operating without a net, and the data shows it responds accordingly.

This isn't a flaw in either platform. It's by design. But it means brands need to understand not just whether they're being cited by AI, but where that citation is appearing and what the platform is being asked to do on its own.

 

AIO Cites Retailers Directly at Twice the Rate of ChatGPT

The most direct expression of this architectural difference: where citations actually go.

In Google AI Overviews, 30% of transactional citations reference a major retailer domain directly. In ChatGPT, that figure drops to 15%. Same purchase-intent query. Half the direct retailer presence.

The gap reflects the division of labor on each platform. AIO doesn't need to do as much evaluative work before pointing to a retailer because the SERP around it provides the commercial context. ChatGPT, operating without that context, routes citations differently before it arrives at a brand recommendation.

For retailers, this has a concrete implication. Being cited in AIO on transactional queries is a different kind of win than being cited in ChatGPT. AIO citation puts you on a page where the user is already in purchase mode. ChatGPT citation puts you in a response that still has more work to do before the user acts.

 

AIO Leans on Social Proof. ChatGPT Doesn't.

One of the more striking findings in the data is how differently each platform uses social and community content to anchor purchase recommendations.

YouTube and Facebook together account for nearly 13% of AIO's transactional citations. ChatGPT surfaces that same category at just 3%. A 4x gap. When Google's AI wants to validate a purchase recommendation, it reaches for peer content: video reviews, community discussions, social proof from real users. ChatGPT largely doesn't follow the same pattern.

This reflects a broader dynamic in how AIO handles the consideration layer. Where ChatGPT needs to do its own evaluative work in the text of the response, AIO can point users toward community content that carries that evaluation implicitly. A YouTube review, a community discussion, a video comparison — these are the validation signals AIO leans on. ChatGPT builds its own.

For brands, this matters beyond citation strategy. If your category's purchase journey is anchored in peer validation, and most retail categories are, your presence in social and video content isn't just a community play. It's an AIO citation surface.

 

ChatGPT Adds a Verification Layer Before It Recommends

Where AIO routes transactional citations toward retailers and social content, ChatGPT takes a different path. It goes to editorial and financial sources first.

Four of the top six most-cited domains in ChatGPT's transactional responses are editorial or financial sources: review outlets, deal-analysis sites, financial comparison platforms. In AIO, four of the top six are retailers. ChatGPT is adding a verification step that AIO largely doesn't need, because AIO's SERP already provides it through organic results and Shopping units.

The implication for brands is significant. A brand that isn't referenced by the editorial and financial sources ChatGPT trusts may be getting filtered out before the recommendation is made. The citation isn't just about whether your domain appears. It's about whether the sources ChatGPT relies on to validate purchases are already vouching for you.

 

ChatGPT Surfaces Wider Competitive Sets

ChatGPT surfaces an average of 7.5 brands per transactional response. AIO surfaces 6.1. That gap compounds across the buyer journey.

More brands per response means more options presented before a decision gets made. For any individual brand, it means the consideration set is wider and the path from AI response to purchase action is longer and more competitive on ChatGPT than on AIO.

This pattern is consistent with ChatGPT's role as a stand-alone evaluative layer. Without the SERP infrastructure to narrow the field, ChatGPT presents the user with more options and more context, letting the response do the comparison work that a SERP might distribute across multiple surfaces.

 

ChatGPT Is More Willing to Surface Negative Sentiment

Negative brand mentions in ChatGPT's transactional responses run at nearly double the rate of AIO's. 0.7% vs. 0.4%.

The absolute numbers are small. But the pattern matters. When ChatGPT is the only thing on the page, it bears full responsibility for a complete and balanced answer. That means it's more willing to surface reasons not to choose a brand, including compatibility issues, price concerns, and product limitations, as part of making its response useful. AIO, operating within a SERP that gives users more ways to evaluate on their own, applies a lighter editorial hand.

The practical implication: brands with product or experience weaknesses that are well-documented in editorial and review sources face more exposure in ChatGPT's transactional responses than in AIO's. Monitoring sentiment in AI responses isn't just a brand exercise. It's a transactional visibility issue.

 

What Marketers Need to Know

The behavior difference is architectural, not algorithmic. AIO and ChatGPT are both trying to answer the same question. The path they take depends on what's around them. Understanding that distinction is the starting point for any AI citation strategy in retail.

Social and video presence is a transactional citation surface on AIO. AIO's 4x higher rate of social and community citations on transactional queries means peer content, including YouTube reviews, community discussions, and video comparisons, is doing citation work in the purchase journey. Brands that don't show up in that content layer are absent at a critical moment.

ChatGPT's verification layer is the editorial web. If the review outlets, comparison sites, and financial sources ChatGPT trusts aren't vouching for your brand, you may be getting filtered before the recommendation is made. Visibility in those sources isn't just an SEO play. It's a ChatGPT transactional citation play.

The good news: the foundation is the same across both platforms. Authoritative content. Trusted source signals. Credibility at scale. The inputs that drive citation visibility on AIO are the same inputs that drive it on ChatGPT. They're just weighted and expressed differently depending on the environment. Brands that build that foundation don't need separate strategies for each platform. They need the visibility to see how each platform is interpreting what they've already built.

 

Technical Methodology

ParameterDetail
Data SourceBrightEdge AI Hyper Cube
Engines AnalyzedGoogle AI Overviews, ChatGPT
Query SetTens of thousands of prompts where top U.S. retailers were mentioned or cited as a source, filtered to transactional intent
Intent ClassificationEach prompt categorized as Informational, Consideration, Branded Intent, Transactional, or Post Purchase
Citation ClassificationCited domains categorized by type: major retailer, social/community, editorial/financial, news media, government/academic, other/niche
Sentiment AnalysisBrand mentions extracted and classified as positive, neutral, or negative across both platforms
Cross-Platform ComparisonHead-to-head citation source and sentiment analysis across both engines using matched query methodologies

 

Key Takeaways

FindingDetail
2x Retailer Citation Gap30% of AIO's transactional citations go directly to a major retailer. ChatGPT: 15%. Same purchase intent. Half the direct retailer presence.
AIO's Social Proof Signal is 4x StrongerYouTube and Facebook combine for nearly 13% of AIO's transactional citations. ChatGPT surfaces that same category at 3%.
ChatGPT Routes Through Editorial First4 of the top 6 most-cited domains in ChatGPT's transactional responses are editorial or financial sources. In AIO, 4 of the top 6 are retailers.
ChatGPT Surfaces More CompetitorsChatGPT averages 7.5 brand mentions per transactional response vs. 6.1 for AIO. Wider competitive sets mean longer paths to a decision.
ChatGPT Carries Nearly 2x the Negative SentimentNegative brand mentions run at 0.7% in ChatGPT's transactional responses vs. 0.4% in AIO. When the AI is the whole page, it does more of the evaluative work, including the critical part.
The Foundation Is the SameAuthoritative content and trusted source signals drive citation visibility on both platforms. The difference is how each environment expresses them, not what builds them.

Download the Full Report

Download the full AI Search Report — How Google AI Overviews and ChatGPT Cite Retailers Differently

Click the button above to download the full report in PDF format.

Published on  March 27, 2026

Optimizing for AI Agents: What Marketers Need to Know About Crawl Behavior

Ensure your content is discoverable, usable, and preferred in AI-powered search experiences

This session goes live on Wednesday, April 15, 2026, at 9:00 AM PDT | 12 PM EDT . Reserve your spot and get ahead of how AI agents are transforming the way your content is found and used.

Description

AI agents are no longer a future concept. They are actively crawling your website right now. From ChatGPT's operator agents to Perplexity, Google's Gemini agents, and beyond, a new class of automated visitors is determining how your brand shows up in agentic search experiences. And they don't behave like traditional search bots. But is your site technically prepared for how these agents access, parse, and act on your content? 

Join us for this technical deep dive into AI agent crawl behavior and what it means for your marketing strategy. While foundational SEO practices still apply, the agentic era introduces a new layer of considerations — and knowing how to communicate their business impact internally is critical for securing the resources to act on them. We'll walk you through what you need to know to ensure your content is accessible, actionable, and preferred by AI agents across the search landscape.

What You'll Learn

  • How AI agent crawlers differ from traditional search bots — and why it matters
  • The technical factors that influence whether agents can access and use your content
  • How to audit your site for AI agent readiness
  • How to make the internal case for prioritizing agent optimization

Featured Speakers:

Dave McAnally
 
Having trouble with the form? Please try refreshing the page or contact us directly at support@brightedge.com

Save your spot 

* indicates required
 
 
 

 
 

How AI Agents Define Your Brand’s Image

English, British
News Item Title
How AI Agents Define Your Brand’s Image
News Item Author Name
Forbes (CMO Network)
News Item Published Date
News Item Summary

Forbes CMO Network explored how AI agents are shaping brand perception as users increasingly rely on generative search and assistants for decision-making. BrightEdge CEO Jim Yu was cited, building on themes from the SPARK Live keynote on March 12, with references to AI HyperCube (AIHC) and AI Agent Analytics in understanding how brands appear across AI-driven experiences. The article highlights how AI-native visibility and measurement are becoming central to managing brand presence in search.

ChatGPT More Likely To Criticize Brands Near Purchase

English, British
News Item Title
ChatGPT More Likely To Criticize Brands Near Purchase
News Item Author Name
MediaPost
News Item Published Date
News Item Summary

MediaPost reported on how AI platforms influence purchase decisions, focusing on how brand sentiment shifts closer to conversion. BrightEdge research was cited showing ChatGPT is more likely to surface critical brand perspectives near purchase-stage queries, based on analysis of AI-generated responses. The coverage highlights how AI-driven search is shaping late-stage decision making and changing how brands are evaluated.

AI Is The Latest Gatekeeper Between Brands And Buyers

English, British
News Item Title
AI Is The Latest Gatekeeper Between Brands And Buyers
News Item Author Name
Semafor
News Item Published Date
News Item Summary

Semafor examined how AI platforms are increasingly shaping purchasing decisions, with brands now optimizing content for AI-driven recommendations rather than traditional search. BrightEdge research was cited, showing Google AI Overviews are more likely to surface negative brand sentiment compared to ChatGPT, based on large-scale analysis of AI-generated results. The article highlights how AI systems are becoming a critical layer between brands and consumers, changing how visibility and influence are determined.

,