SEO Myths that Circulated to SEO Beginners Part Two
# 3: (XML) Sitemaps Are Necessary for a Good Placement!
Google always knows exactly what it can find in the respective web project, right? Well, it's not!
In fact, HTML and XML sitemaps are extremely important - for the actual indexing. A good website will therefore contain an appropriately prepared sitemap - as well as a robots.txt file with instructions for search engine crawlers and bots. Both are helpful for search engines to index the website correctly.
However, the sitemap has no influence on the placement within the search results.
# 4: Duplicate Content (DC) is Evil!
Duplicate content leads to penalties and should therefore be avoided in any case.
As a rule, Google recognizes whether the content is identical under different URLs - and since when it has been listed in the index.
Correspondingly, older, trustworthy domains are usually recognized as the origin and not penalized.
In contrast, scraping pages are usually very quickly identified - and punished - due to their short domain age, the simultaneous mass absorption of already known content, and the missing backlinks.
Duplicate content on different domains that are created intentionally and inevitably should be well marked:
Set canonical tags
Set references/links for users
This applies, for example, to news and press articles, specific manufacturer information in eCommerce projects, or articles that are deliberately reposted in various other places.
DC unavoidable under the same URL is usually created in eCommerce projects, catalogs, or the like.
The important thing here is toavoid unnecessary duplication
Keeping necessary structural duplications out of the index (declaration via robots.txt file)