The End of &num=100: Is Google Fortifying its Ecosystem?

Insights from our Research Desk

In mid-September 2025, Google quietly removed support for the &num=100 parameter, which had allowed users to view up to 100 search results on a single page instead of the default 10. The feature was a staple for researchers, SEO analysts, and developers who relied on it to collect and analyze larger datasets efficiently. For example, an SEO analyst at a marketing agency might use &num=100 when researching how a client’s website ranks for a keyword like “activant research” (e.g., https://www.google.com/search?q=activant+research&num=100). By seeing all 100 search results at once, they can quickly scan competitors, identify trends in top-ranking pages, and export the full list for analysis without clicking through multiple pages.

While the shift affects reporting rather than rankings, it disrupts a decade of SEO trendlines and makes large-scale scraping materially harder. Expect cleaner Google Search Console metrics, higher costs for scrapers and agents, and more control in Google’s hands.

The real question is whether this was a direct response to the DOJ’s antitrust ruling and a pre-remedy move to fortify Google’s position ahead of court-mandated data sharing.

What changed?

From September 10–14, 2025, industry data shows widespread drops in Search Console impressions and keyword counts, while clicks and actual rankings remain largely unchanged. This timing aligns with the disablement of the &num=100 parameter, suggesting the decline reflects reduced query-level visibility and data availability rather than real traffic loss or ranking shifts. This means:

  1. Cleaner data: Fewer “deep-page” bot impressions are counted, lowering total impressions but raising average positions since lower-ranking pages no longer dilute results. Click-through rates now better reflect genuine user behavior.

  2. Higher scraping costs: Rank trackers and AI agents must now paginate, issuing 10X more requests and hitting rate-limits and CAPTCHAs more frequently.

  3. Measurement: For over a decade, tools supporting visibility indices, rank tracking and market share modeling relied on the &num=100 parameter. Its removal disrupts historical trendlines and breaks backward comparability.

Was this strategic?

Although Google has not made a formal announcement, the quiet removal of the parameter appears deliberate. The move is best understood not as a single action, but as a calculated step in a much larger strategy involving data control, antitrust pressure, and the future of generative AI.

Raising the Walls Against Scrapers

The immediate effect of disabling &num=100 is to restrict bulk retrieval of search results that have long powered SEO tools, competitive intelligence platforms, and AI training pipelines. Historically, scraped result pages have fueled everything from market share trackers to retrieval-augmented generation (RAG) systems. Because Google offers no official Search API with full access to live results, third-party services like SerpAPI emerged to provide this data programmatically. Last year, it was reported that AI firms like OpenAI and Perplexity relied on SerpAPI to fetch live results for real-time answers. This change aligns with Google’s broader anti-scraping campaign, which includes hiring for an “Anti-Scraping Engineering Analyst”, deploying tighter CAPTCHAs, and improving bot detection. Together, these actions point to a concerted effort to raise the cost of large-scale data harvesting and preserve the exclusivity of its search data.

Pre-Positioning for Antitrust Remedies

This defensive move appears to be part of Google’s pre-remedy positioning ahead of court-mandated data-sharing measures stemming from the DOJ’s antitrust case. The recent ruling requires Google to provide rivals with certain search index and user-interaction data (including a one-time snapshot of URLs and webpages it has indexed). Furthermore, Google must operate a temporary syndication program to help new entrants cover long-tail and local queries while they build their own indexes.

These long-tail queries are rare, specific searches that make up most query volume, and illustrate the industry’s “80/20 problem”. While covering the first 80% of common queries is relatively achievable, the remaining 20% requires massive, constantly refreshed data and large-scale click-feedback. Court testimony from Microsoft and DuckDuckGo emphasized that this user-side and tail-query data is critical for any competitor to close the gap, and that LLMs alone cannot substitute for it.

Cementing Control of the GenAI Data Layer

Google’s vast index gives it a clear and powerful advantage in handling these long-tail queries. The strategic importance of this data was further underscored in testimony from Microsoft and OpenAI, which argued that limited access to high-quality search APIs is an “innovation killer” for GenAI products. OpenAI explicitly testified that access to Google’s API would materially improve its models’ reliability and user experience.

Given these coming regulatory requirements, curbing unauthorized scraping activity makes perfect strategic sense. Disabling num=100 raises the cost of bulk data extraction, preserving Google’s long-tail data advantage until formal data-sharing channels take effect. This effectively steers competitors away from unauthorized scrapers and toward regulated, auditable data sources. If these new remedies succeed, smaller General Search Engines (GSEs) and GenAI entrants could use Google’s index breadth and click signals to narrow the 80/20 gap in search coverage; if they falter, Google’s integration of AI Overviews (AIO) with Search will only reinforce its current data lead.

Implications for Founders (especially AI and data products)

  • RAG/agents: Budget for higher retrieval latency and cost if you depend on live SERPs. Add caching, batch queries, and consider licensed feeds or partnerships.

  • Training data: Expect shallower long-tail coverage from naive scraping, skewing toward mainstream domains unless you diversify sources. The resulting bias favors established publishers and commercial domains.

  • SEO Measurement: Reset baselines from Sept 12, 2025. Compare clicks and conversions, and annotate the impressions break in all dashboards.

  • Go-to-market: If your product resells search-derived metrics (rank, visibility, share of voice), communicate the methodology change and re-benchmark. SISTRIX, among others, has already adjusted collection methods.

  • Enterprise Search: Expect differentiation to hinge on proprietary data access, domain-specific embeddings, and user-context integration. The moat lies in owning structured, permissioned corpora (emails, docs, tickets, CRM data), building deep connectors into enterprise SaaS, and maintaining compliance-grade retrieval pipelines that cannot be replicated through public web data.

Bottom Line

With a six-year remedy period, Google has time to shape implementation, entrench its ecosystem, and attempt to adapt compliance on its own terms. Google is narrowing automated access to the discovery layer. The move reflects a deliberate pre-remedy posture designed to fortify Google’s AI ecosystem ahead of regulatory enforcement and cement its control over the foundational data layer of the GenAI stack. Without robust search APIs, competing GenAI assistants will struggle to retrieve real-time information, cementing Google’s edge in a market where grounded reasoning and real-time accuracy are defining advantages. For AI startups, the smart move is to de-risk from raw SERP scraping: shift to licensed sources, strengthen retrieval ops, and invest in first-party data. For everyone else, mark mid-Sept ’25 as the break in your time series and keep shipping.