Deep Link Requirements Tighten

Google has clarified that to qualify for “Read more” deep links in search results, content must be immediately rendered on page load. Sites relying on JavaScript-heavy loading, tabbed interfaces, or hidden sections will be excluded. This change forces engineering teams to prioritize raw HTML accessibility over complex client-side rendering to maintain search visibility.

Robots.txt Standardization

Google is formalizing documentation for common “unsupported” rules in robots.txt files. By integrating support for the most frequent developer typos and edge-case directives, Google is reducing the technical overhead for site maintenance while signaling a shift toward more predictable crawler behavior.

The EU’s Data Mandate

The European Commission has issued preliminary findings requiring Google to share core search dataโ€”including query, click, and ranking dataโ€”with competitors and AI search providers under the Digital Markets Act (DMA). If enacted, this breaks Google’s primary data moat, enabling third-party search engines and AI wrappers to train models and optimize rankings on Google-grade data.

Implications

For operators, the SEO updates are tactical: audit your landing pages immediately to ensure “Read more” content is rendered server-side. The engineering cost of refactoring client-side scroll or hash-based navigation is now a direct tax on your organic acquisition funnel.

The EU’s proposal represents a structural threat to Google’s search monopoly. If search data becomes a commodity on FRAND terms, the competitive advantage of proprietary ranking algorithms diminishes. Founders building in the search or AI space should prepare for a landscape where ranking data is no longer exclusive to the incumbent, potentially accelerating the development of niche, intent-specific search engines.