About 1 month ago I encountered this graph in a Reddit post, which starkly illustrated the DDoS level of traffic that AI can generate when it's mining a site.

Yes it's concerning, but it also inspired some intriguing questions...
- How much AI traffic is going directly to the HTML pages on my Webflow site?
- How much of that traffic is invisible to my traditional analytics tools?
- Where is that traffic coming from?
- What are those AI bots doing?
And, of course... questions demand answers.
The Experiment
10 days ago I set up a reverse proxy on Sygnal's site to log all AI traffic it could identify.
Since AI bots do not always announce themselves clearly in the User Agent string, I took a more thorough approach using about 20 published CIDR lists. If you're unfamiliar with the term, CIDRs are are IP-address ranges which are used to identify a particular service through the IP's it uses.
With that approach I could directly identify OpenAI, Perplexity Google, and even the specific bots from each that are visiting.
I launched that on Oct 13, 2025, curious if anything might show up.
And show up, it did.
How much AI traffic is going directly to the HTML pages on my Webflow site?
Certainly more than I expected.

The brown line at the bottom that starts on Oct 12th represents normal pageviews. Everything else is AI traffic, each line representing a different bot and provider.
Where is that traffic coming from?
Here's that list.

A few important things to notice here;
- The wide range of different bots. OpenAI for example, has 3 different bots which each serve a different purpose.
- The bot type. This is a classification I've assigned to help differentiate bot roles.
How much of that traffic is invisible to my traditional analytics tools?
This is somewhat difficult to answer, but the broad answer is "most of it."
Bots which do not execute JavaScript, and which do not load referenced files like pixels won't be detected by traditional analytics like GA4.
However that does appear to be changing.
Here are Sygnal's 2024 GA4 user activity patterns from Jan - Dec 2024. Fairly smooth, and organically tied to seasons, weather and weekly traffic cycles.

Here's the same report for this site for 2025, dating Jan 2025 to today, 23-Oct-2025.
It looks a lot less... natural.

What are those AI bots doing?
Now this is interesting. I asked ChatGPT to analyze the table and to tell me what it could about each of the bots I'm seeing.
Here's what it reported.
OpenAI (ChatGPT-related)
Provider | What It Is | Purpose | Type |
---|---|---|---|
openai-chatgpt-user | ChatGPT User Agent | Triggered when a human user inside ChatGPT browses your site via the "Browse with Bing" or integrated browsing features. | Assistant (human-triggered) |
openai-gpt | GPTBot | OpenAI's main web crawler used to collect publicly available content to improve AI models. | Crawler (AI training) |
openai-search | OpenAI SearchBot | Used for retrieval-augmented responses in ChatGPT (especially when browsing, RAG queries are needed). More targeted than GPTBot. | Crawler (search, RAG) |
Microsoft (Bing + ChatGPT integration)
Provider | What It Is | Purpose | Type |
---|---|---|---|
microsoft-bingbot | Bingbot | Bing's traditional web indexer, also used indirectly by ChatGPT browsing when Bing integration is used. | Crawler (search engine index) |
Provider | What It Is | Purpose | Type |
---|---|---|---|
google-googlebot | Googlebot | Standard Google search index crawler. | Crawler (search) |
google-user-fetchers-google | Google "user-triggered fetchers" | Usually fired by features like Google Web Preview, Lighthouse testing, or Google-generated link fetches due to user actions. | Crawler (on-demand fetch) |
google-special-crawlers | Google special-purpose crawlers | Used by specific services e.g. Google AdsBot, Google-Image, Google-Site-Verification, etc. | Crawler (service-specific) |
Perplexity AI
Provider | What It Is | Purpose | Type |
---|---|---|---|
perplexity-bot | PerplexityBot | Crawls content to power Perplexity AI's large-language-model search engine. | Crawler (AI training/search) |
perplexity-user | Perplexity User-Triggered | Represents real user queries executed via Perplexity's assistant interface (like ChatGPT's browsing). | Assistant (human-triggered) |
DuckDuckGo + AI Search
Provider | What It Is | Purpose | Type |
---|---|---|---|
duckduckgo-bot | DuckDuckBot | General DuckDuckGo search crawler. Likely now feeding DuckDuckGo's AI summaries. | Crawler (search/AI) |
Common Crawl (used by many AI models)
Provider | What It Is | Purpose | Type |
---|---|---|---|
commoncrawl-ccbot | CCBot | Collects massive open web datasets used for AI training (OpenAI, Meta, AWS Bedrock, etc. use Common Crawl). | Crawler (AI dataset provider) |
Summary by Usage Category
Category | Agents |
---|---|
🔍 General Search Indexers | bingbot , googlebot , duckduckbot |
🤖 AI Training Crawlers | openai-gpt , perplexity-bot , commoncrawl-ccbot |
📚 AI Retrieval/Search (RAG) Bots | openai-search , likely used directly for query-time data |
👤 User-Agent AI Requests (human-triggered) | openai-chatgpt-user , perplexity-user , google-user-fetchers-google |
🛠️ Service-specific Crawlers | google-special-crawlers |
- Assistant-type (human-triggered) requests usually reflect real user intent and are valuable.
- Crawler-type (AI training or indexing) may be fine, but you may want to:
What I've Learned
AIO is here, and it's probably all that matters now.
Basic SEO likely matters as well, but in my opinion, technical SEO is likely to be a wasted of effort. LLMs don't need JSON-LD or META tags to understand and use your content.
Page titles, paths, and headings probably matter- but more for Google's benefit. Having your site appear in search engines will still have value in 2026, mostly because LLMs will look there as well.
Traditional Analytics like GA4 are rapidly losing value
Unless AI bots start running JavaScript, using cookies, downloading full page content, they'll be invisible to client-side analytics pixels.
Even if that did happen, bots aren't likely to discover or navigate a site like user sill because they have a billion bookmarks that take them straight to the content they want... which means traffic flow analytics and funnels are broken.
The lack of mouse movements, scroll distance, click events, means session recordings are becoming less useful.
Bots are not the same
And even bots within our contrived "Assistant" category behave very differently in terms of how they use and convey your information.
Traffic measurement needs a serious rethink
Sessions and user flows are now useless, because AI bots don't approach your site in the same way... and it's likely they don't use cookies at all.
Conversion approach needs a serious rethink
Customer-acquisition funnels are probably dead for the same reason- it's not how customer will discover, evaluate, or decide on your product. The LLM becomes the funnel, and you have a lot less control there.
Upsell funnels are still 100% strong because once you have the customer's details, you can contact them directly without LLM interference.
Rethinking the Approach
Currently I'm thinking that a more useful analysis would fall along two lines;
Attention ( aka signal strength )
Historically, someone saw your page, or they didn't. Attention was binary- they got 0% of your signal, or 100%.
And if I read my above traffic report with that mentality, I see this.
Which would suck.

But that's not accurate, because a good portion of that AI traffic human proximate. A person was nearby, and was delivered my information, hopefully with fewer typos.
Let's assume ChatGPT described assistant bots accurately. That means that my signal strength isn't limited to the 1.136 page views this week. It also includes the assistant AI traffic, another 7,898 page views.

Let's take that idea further...
Now that AI's filter and meld your content in multiple ways, signal strength depends on the context in which it was delivered.
Signal strength | Notes | |
Model | 1% - 10% | Occurs when your information has been fully absorbed into the model, and blended with 1,000's of other sources. Since models are expensive to build, your information can be a year or more out of date as well. |
RAG | 20% - 70% | Occurs when the LLM performs realtime research, usually by querying Google, and finds and parses your site in realtime. The information is current, but it will be integrated into a response with other sources as well. |
Assistant | 80% - 100% | Some browsers like Bing and OpenAI's new browser seem to be operating as "augmented search" Your content might be diluted or interspersed with other content but it still takes a front-stage role in the response. |
Link | 100% | Occurs when the LLM response includes a link directly to your site. Of course, that link must be current, not 404, and the user must click it. |
I like this perspective. It feels better aligned with today's reality.
Engagement ( conversions )
The end goal is still key. But we have new challenges.
Does the user buy? Do they register? Do they learn? Do they like & follow?
The challenge now is that we lose visibility into that path, and have a lot less control over how information, branding, and value proposition are presented.
If a user is trying to solve X, or find product Y using an LLM, new questions arise.
- Discovery - Were they made aware of our product / service
- Presentation - How was it presented alongside other options? Was it presented positively or negatively?
- Evaluation - Did the LLM "sell" it well as a viable solution and give the correct, critical decision
- Decision - What did the decision come down to as the user narrowed their choices and selected or abandoned your offering.
- Action - Then what? Was a link provided? Which link? How does the user convert?
- CRO - When we get a conversion, how do we trace the path back to the beginning? At this point, it's fairly opaque.
A New Spin on an Old Problem
If you're seeing this the same way I am, it's the same problem that broadcasters had when they released new TV shows. "Is anyone watching this?" was a multimillion-dollar question for every show, and every commercial.
Companies like AC Nielsen stepped in to fill that gap by building track & measure systems at the consumer end, to monitor and record consumer behavior. They reported, and sold- the who, the what, and the when.
I think soon enough, we'll see that here too- probably in the form of AI companies themselves using the free services to capture user data, and monitor and track their interactions all the way through to a sale or conversion event. Then sell that data back to large companies and ad agencies.
Instead of digging through analytics and looking at recorded sessions, you'll be reading anonymous, redacted LLM discussions where someone was trying to decide on the right new phone to buy.
Interesting times ahead, I guess.