Bug #89617
closedXML sitemaps should not be indexed by search engines
100%
Description
The content of a XML sitemap should not be listed on search engine result
pages. This is confirmed by John Mueller (Webmaster Trends Analyst from Google):
Updated by Gerrit Code Review about 5 years ago
- Status changed from New to Under Review
Patch set 1 for branch master of project Packages/TYPO3.CMS has been pushed to the review server.
It is available at https://review.typo3.org/c/Packages/TYPO3.CMS/+/62253
Updated by Gerrit Code Review about 5 years ago
Patch set 2 for branch master of project Packages/TYPO3.CMS has been pushed to the review server.
It is available at https://review.typo3.org/c/Packages/TYPO3.CMS/+/62253
Updated by Gerrit Code Review about 5 years ago
Patch set 1 for branch 9.5 of project Packages/TYPO3.CMS has been pushed to the review server.
It is available at https://review.typo3.org/c/Packages/TYPO3.CMS/+/62219
Updated by Chris Müller about 5 years ago
- Status changed from Under Review to Resolved
- % Done changed from 0 to 100
Applied in changeset cdab79385c12d4d4bf4059f184243a93c1d4caf3.
Updated by Gerrit Code Review about 5 years ago
- Status changed from Resolved to Under Review
Patch set 2 for branch 9.5 of project Packages/TYPO3.CMS has been pushed to the review server.
It is available at https://review.typo3.org/c/Packages/TYPO3.CMS/+/62219
Updated by Chris Müller about 5 years ago
- Status changed from Under Review to Resolved
Applied in changeset ff6c916fbd63e90d714a89ea58f99ea567ced14a.
Updated by Mordamir over 2 years ago
Whats the reason behind this decision?
With the x-robots-tag set to noindex, google search console does not accept the sitemap. And a xml sitemap is mostly for crawlers. So why should i exclude the crawlers from using the sitemap?
Maybe this is recent change from google?
Updated by Chris Müller over 2 years ago
"With the x-robots-tag set to noindex, google search console does not accept the sitemap."
I can't reproduce that. Added some sitemaps to GSC which is possible some time ago with this change.
You have to distinguish between crawling and indexing. Crawling means, that the crawler is able and allowed to retrieve the content and follow the links. This is possible for XML sitemaps with the "X-Robots-Tag: noindex" header. This is the same for a meta tag "robots" with "noindex" on a page: The page is crawled (and links followed) but the page isn't indexed (= displayed in the search results).
But: The XML sitemap should not be visible in the results of a search engine. Thus the "X-Robots-Tag: noindex". See also the link in the description of this issue.
And: If you want to allow the indexing of the XML sitemap (which means displaying in the search results) for whatever reason, you can easily achieve this by adding the following line in the TypoScript setup of your site package:
seo_sitemap.config.additionalHeaders.20 >
Updated by Florian Seirer almost 2 years ago
Had the same issue as @Mordamir : Our XML sitemaps were not crawled by Google for weeks because of the noindex header, according to the report in the Google Search Console.
But simply resubmitting the sitemaps fixed the problem. All URLs from the sitemaps are now recognized.
Maybe this was just a temporary issue on Google's side?