Back to blog
|6 min read

GA4 Bot Traffic: How to Detect and Filter It

Not all traffic in GA4 comes from real users. Bots, crawlers, and automated scripts account for a significant share of web traffic on most properties, and unless you actively filter them, they inflate session counts, distort engagement metrics, and create phantom conversion signals.

GA4's Built-In Bot Filtering

GA4 includes automatic bot filtering that uses Google's internal list of known bots and spiders, sourced in part from the IAB/ABC International Spiders and Bots List.

This filtering is enabled by default for all GA4 properties and removes traffic from identified bots before it enters your reports. However, this filtering is far from comprehensive.

It catches well known, self identifying crawlers, legitimate search engine bots, monitoring tools, and automated testing services that follow standard user agent conventions.

It does not catch sophisticated bots that mimic human browser behaviour, use residential proxy networks, or deliberately impersonate real user agents.

For properties with significant traffic, it is worth verifying that bot filtering is enabled in the data stream settings (under "Advanced settings" in each data stream) and supplementing it with manual analysis to catch what the automated filter misses.

Identifying Bot Traffic Patterns Manually

Manual bot detection requires looking for patterns that are statistically unlikely for human users. Unusually high session counts from a single country or city that does not match your target audience is a common signal.

Engagement rates near zero across many sessions, every session ending in under one second, indicates non human traffic executing a page load without rendering or interacting.

High event volumes with no conversion activity can indicate scrapers or monitoring bots.

Hostname anomalies are another indicator: if your GA4 data shows sessions where the hostname does not match your actual domain, those are likely bot sessions originating from tag injections or referral spam.

Look at the device and browser distribution too, legitimate traffic shows natural diversity, while bot traffic often clusters around a single browser version or device category.

Configuring Filters and Internal Traffic Rules

For bot traffic that GA4's automatic filtering misses, create data filters in the GA4 property admin. You can create IP-based filters to exclude known bot source ranges identified through server log analysis.

For bots that are harder to identify by IP, consider adding a JavaScript check in your tag manager that verifies the browser environment looks legitimate before firing the GA4 tag, checking for headless browser indicators like navigator.

webdriver being true is a basic but effective signal. Note that GA4 data filters only apply to new data going forward and cannot be retroactively applied to historical data, so identifying and configuring filters promptly minimises the period of contaminated data in your reports.

Ready to audit your GA4 property?

Run a full GA4 audit in under 10 minutes. Free to start.

Start Free Audit