What does Google say about SEO? /

Official statement

To implement structured data markup at scale, the best approach is to do it programmatically via the CMS or developers, rather than manually on a page-by-page basis. Google Tag Manager is an alternative but is more fragile.
31:49
🎥 Source video

Extracted from a Google Search Central video

⏱ 38:29 💬 EN 📅 18/05/2020 ✂ 10 statements
Watch on YouTube (31:49) →
Other statements from this video 9
  1. 1:06 Is dynamic rendering really risk-free for SEO?
  2. 1:38 Does dynamic rendering really slow down your server or enhance your crawl budget?
  3. 2:39 Why does Google treat JavaScript redirects as 302s instead of 301s?
  4. 2:39 Does Google really treat 301 and 302 redirects differently for SEO?
  5. 3:42 Can Googlebot really crawl hidden links in a hamburger menu?
  6. 5:46 Should You Serve Lightweight Pages to Bots to Enhance Performance?
  7. 7:01 How can you effectively manage 404 errors in a SPA without risking deindexation?
  8. 14:57 Why is Googlebot missing your content loaded by Web Workers?
  9. 30:51 Is it true that hidden content in accordions is actually indexed by Google?
📅
Official statement from (5 years ago)
TL;DR

Martin Splitt claims that programmatic implementation of structured data via CMS or developers remains the most reliable method for deploying large-scale markup. Google Tag Manager, while convenient, has structural weaknesses that may compromise data reading by search engines. For a medium to large site, this technical approach becomes essential if you aim for comprehensive and stable coverage over time.

What you need to understand

Why does Google recommend a programmatic approach over a manual one?

The answer boils down to one word: scalability. On a site with 50 pages, manually editing each structured data block is manageable. But once you exceed a few hundred pages — or worse, thousands — maintenance becomes a logistical nightmare.

Programmatic integration via the CMS or server-side templates ensures that each new page automatically inherits the correct schema. A new product added? The template generates the Product schema. An article published? Article schema deployed. Zero manual intervention, zero omissions.

Google values this method because it drastically reduces the error rate. The tags are consistent, the required properties are present, and updates propagate with a single code change. This is exactly what a search engine is looking for: structural consistency at scale.

What’s the difference between server implementation and Google Tag Manager?

Server-side implementation injects JSON-LD directly into the initial HTML returned to the browser. Googlebot sees the structured data from the first render, without needing to execute JavaScript. It’s fast, reliable, and completely independent of the client environment.

Google Tag Manager, on the other hand, loads structured data after the page has loaded through JavaScript. This means that if GTM takes time to execute — due to slow connections, blocking scripts, or JS errors — the data may arrive too late or may never be read. Splitt calls this method "more fragile," and the evidence shows this: cases regularly occur where Google does not detect markup injected via GTM, particularly on mobile or with limited crawl budget.

In what contexts does programmatic implementation become essential?

As soon as your site includes dynamically generated content — e-commerce, classified ads, marketplaces, aggregators — you have no choice. No one is going to manually edit the schema of 10,000 product listings that change prices daily.

Editorial sites with high publication volume are in the same boat. If you publish 20 articles a day, you want each writer to focus on their content, not on JSON-LD syntax. The CMS should handle this behind the scenes, leveraging already entered metadata (author, date, category, featured image).

  • Scalability: automatic deployment across thousands of pages with no manual intervention
  • Reliability: immediate readability by Googlebot, no JavaScript dependency
  • Maintenance: a single template modification updates the entire site
  • Consistency: uniformity of schemas, reduction of error rate
  • Crawl performance: data available from the first HTML render

SEO Expert opinion

Is this recommendation really absolute or are there exceptions?

Splitt is correct in principle, but the field reality is more nuanced. If you manage a small showcase site of 30 static pages that only changes once a quarter, manual implementation remains perfectly viable. You control exactly what is published, can test each schema individually, and don’t need to mobilize a developer.

The real threshold is around 100-150 pages with regular updates. Below this, the cost of programmatic integration might exceed the benefit. Beyond that, you lose too much time and take too many risks of inconsistency. [To be confirmed]: Google publishes no numerical data on the error rate comparison between the two methods, but field experience clearly shows that high-volume sites with manual markup multiply omissions and outdated schemas.

Why is Google Tag Manager considered "more fragile"?

Because GTM introduces a dependency on JavaScript that does not exist with server implementation. If a third-party script fails, if the dataLayer isn't correctly populated, or if Googlebot's rendering budget is exhausted, the structured data may simply never be read.

There are also cases where Google detects the markup but with a delay. The schema appears in the Search Console several weeks after deployment, while a server implementation is visible within 48-72 hours. It’s not a compliance issue — the markup is technically valid — but rather a matter of reading reliability. For a site looking to achieve rich results quickly, this fragility can be costly.

Warning: If you’re already using GTM for structured data and everything is working (active rich results, no errors in Search Console), do not change anything immediately. But if you notice detection issues or plan to scale, migrate to a server-side programmatic implementation.

What hybrid approach can be considered to mitigate risks?

Some sites deploy basic server-side markup (Organization, WebSite, BreadcrumbList) and use GTM for more specific schemas that require behavioral data (AggregateRating from an API, Event with dynamic availability). It’s an acceptable compromise if you manage the dataLayer and actively monitor Search Console.

But let’s be honest: this approach doubles the potential error surface. You have to maintain two systems, debug two sources of markup, and manage potential conflicts between server schemas and GTM schemas. Most sites that test this solution end up migrating everything to server-side once they reach a certain technical maturity.

Practical impact and recommendations

What concrete steps should be taken to migrate to a programmatic implementation?

The first step: audit the existing structured data. List all deployed schemas (Product, Article, Organization, FAQ, etc.), identify those that are manual, those that go through GTM, and those already generated by the CMS. Use a crawler like Screaming Frog or OnCrawl to extract JSON-LD from all your pages and identify inconsistencies.

Next, work with your developers to integrate the generation of structured data into your server-side templates. If you are on WordPress, plugins like Yoast or RankMath do this natively. On Shopify, it’s handled by default for products. For a custom or headless CMS, you will need to code the logic yourself — generally by leveraging already present metadata in the database.

What critical mistakes must be absolutely avoided during migration?

Never remove old markup before the new one is validated in production. First, test on a few pilot pages, verify in Search Console that Google is correctly detecting the new schemas, and then deploy progressively. A massive untested deployment can make all your rich results disappear overnight.

Be careful about duplicates: if you had manual structured data AND a plugin that generates its own, you risk publishing the same schema twice with conflicting properties. Google may then ignore both. Use the rich results testing tool to validate each template before going live.

How can you verify that the programmatic implementation is working correctly?

Crawl your site after deployment and extract JSON-LD from 20-30 representative pages. Check that each type of page (product, article, category) generates the correct schema with all required properties. Compare with Schema.org documentation to ensure compliance.

Monitor Search Console: Improvement section, subsections Product, Article, FAQ, etc. If you see errors appearing after migration, intervene immediately. A malformed schema can block the display of rich results for weeks, while Google recrawls and reevaluates your pages.

  • Audit existing structured data with a crawler (Screaming Frog, OnCrawl)
  • Test programmatic implementation on a subset of pilot pages
  • Validate each template with Google’s rich results testing tool
  • Check for duplicates (old markup + new overlapping markup)
  • Monitor Search Console for 2-4 weeks after full deployment
  • Document generation logic to facilitate future updates
Programmatic implementation of structured data is a technical project that requires good coordination between SEO and developers. If your internal team lacks resources or expertise in these areas, enlisting a specialized SEO agency can save you valuable time and avoid costly errors. Personalized support helps define the right migration strategy, prioritize high-impact schemas, and establish effective monitoring to secure your rich results in the long term.

❓ Frequently Asked Questions

Peut-on utiliser Google Tag Manager pour du structured data sur un gros site e-commerce ?
Techniquement oui, mais Google le déconseille explicitement pour les raisons de fragilité évoquées. Sur un catalogue de plusieurs milliers de produits, l'implémentation serveur reste beaucoup plus fiable et maintenable.
L'implémentation programmatique nécessite-t-elle obligatoirement l'intervention d'un développeur ?
Sur WordPress ou Shopify, des plugins gèrent ça nativement sans code. Sur un CMS custom ou headless, oui, vous aurez besoin d'un développeur pour coder la génération de JSON-LD dans les templates.
Si mon structured data via GTM fonctionne actuellement, dois-je absolument migrer ?
Non, si vos rich results s'affichent et que la Search Console ne remonte aucune erreur, vous pouvez conserver cette implémentation. Mais prévoyez une migration si vous scalez ou si vous constatez des problèmes de détection.
Combien de temps faut-il pour migrer vers une implémentation programmatique ?
Comptez 1-2 semaines pour un site moyen : audit, développement, tests, déploiement progressif. Sur un gros site avec plusieurs types de schémas complexes, prévoyez plutôt 4-6 semaines.
Quelle est la différence de performance crawl entre implémentation serveur et GTM ?
L'implémentation serveur livre le structured data dès le HTML initial, donc lecture immédiate par Googlebot. GTM nécessite l'exécution de JavaScript, ce qui peut retarder ou empêcher la lecture, surtout sur mobile ou avec budget crawl limité.
🏷 Related Topics
Domain Age & History Structured Data AI & SEO Images & Videos Pagination & Structure

🎥 From the same video 9

Other SEO insights extracted from this same Google Search Central video · duration 38 min · published on 18/05/2020

🎥 Watch the full video on YouTube →

Related statements

💬 Comments (0)

Be the first to comment.

2000 characters remaining
🔔

Get real-time analysis of the latest Google SEO declarations

Be the first to know every time a new official Google statement drops — with full expert analysis.

No spam. Unsubscribe in one click.