SEO myths are a dime a dozen. Some of them have their origin in an SEO measure that has long since become obsolete.
Others are based on mere conclusions that have gotten stuck in the minds of SEOs. Even in 2021, there will still be iron SEO myths that are not (no longer) true.
Here are some of the most persistent SEO myths, coupled with official responses from Google reps like John Müller and Matt Cutts.
Mere keyword optimization of the page is no longer enough for a page to be considered relevant.
It is no longer necessary for keyword X to occur a certain percentage in a text.
If only because Google recognizes synonyms and determines the context of the content.
Pages that are not optimized for this keyword but still deal with the topic in a highly relevant manner can rank for a keyword.
Instead of focusing on keyword density, the user (s) should be the focus. Ask yourself:
You can answer all of these questions with the help of a SERP analysis. We have created a step-by-step guide for you that explains exactly how to analyze the search results to create really useful content.
Meta keywords are part of the metadata that every URL has. This data set consists of.
Meta title and meta description are very important parts of every URL. Together with the URL, they result in the so-called snippet, i.e. the search result in the search engines.
Your snippet is your figurehead and has to convince people to come to your page from the search results. The meta title is also a ranking factor.
In order for your search result to be as convincing as possible, you should optimize the meta title and description.
… And that is since at least 2009! Search engines’ language understanding has long been good enough to determine the subject and context of the content.
In the video, Matt Cutts says very clearly: “We don’t use the keywords meta tag for ranking … it was just too much spam. We do not use this information. “
Headings structure an article and subheadings can appear in any number.
BUT: These are only recommendations. Google works with what it finds on the web. As John Müller explains in the video, the web is full of incorrectly used headings.
A highly relevant answer to a search query is not withheld from the user just because there are several H1 headings. And you won’t get yourself a Google penalty either.
You should (as with everything else) worry more about your users than about the search engine.
A meaningful page structure with neatly placed headings will make it easier for them to grasp and understand your content.
This is why SEO tools also output multiple H1 headings, missing H2 headings, or incorrect nesting as errors.
Social signals are not a direct ranking factor on Google. Matt Cutts confirmed this back in 2014 in a video from Google Search Central.
In the video, he says very clearly: “Don’t assume that just because there is a signal from Facebook or Twitter that Google has access to it. Pages can be blocked, there are no-follow links or the like. ”
However, there are correlations between frequently shared content and good rankings:
Like no-follow links, social signals are anything but useless. After all, they also bring important traffic to your site. And: The more often your content is shared via social media, the more people will come to your site, interact with it, and generate positive user signals.
The attention that frequently shared content receives often gives them more backlinks – which in turn are a ranking factor and can bring even more traffic to the site.
The fact that content is often shared on social media is because the content is particularly useful and/or entertaining.
There is no duplicate content penalty. If Google finds identical content several times, it either looks for a variant and ranks it in the search results. Or, there are permanent fluctuations in the rankings, so that a stable ranking cannot be established for a page.
Then one also speaks of keyword cannibalization, triggered by duplicate content.
You see, you don’t need to worry about a Google penalty, but you do about your rankings. It therefore always makes sense to identify and fix internal duplicate content.
Internal duplicate content can have many causes, for example:
You can prevent duplicate content in the search engines by using the canonical tag. The Canonical Tag contains the URL of the “original” and shows the search engines which page should really appear in the search results.
Tips:
Here you can find a detailed article on Duplicate Content and Canonical Tags. You will also learn how to use the canonical tag correctly and what to look out for.
If you find an exact copy of your article or your picture on someone else’s website without ever having given your permission, you don’t have to accept it.
Inform the webmaster (s) of this and ask for immediate removal within a specified period. If he/she does not comply, you can request Google to remove content from Google.
This will not remove your article from the foreign site, but at least it will no longer appear in the search results.
Plus, this is a clear copyright infringement and the law is on your side. If there is no reaction, you can call in the lawyer.
Of course, you should also not copy any content from third-party websites and thus make yourself liable for copyright infringement.
Myths:
Fact:
Crawling does not lead to better rankings – crawling is the technical prerequisite for ranking.
For Google to be able to index and position a page, the search engine must first of all
The crawling allows the search engine to see the website (or at least to see parts of it that it is allowed to see). Which pages she is allowed to look at and which not is controlled by the robots.txt.
The Googlebot looks at the individual sub-pages so that the content can be classified in the next step. So if a page can’t be crawled, it can’t rank either.
How often a website is crawled is determined by the crawl budget allocated to each website.
The aim is for all new features and changes to the website to be included in the search engine index (SERPs) as quickly as possible and to achieve good or better rankings in the second step.
Crawl budget optimization is particularly relevant for large websites with many subpages, such as online shops.
The optimization of the crawl budget is not a ranking factor and does not lead to better rankings.
But it is a measure to get new and updated content in the index more quickly – and thus to achieve better rankings more quickly.
The crawl budget quickly becomes tight for large websites that make many changes every day or that have a large number of subpages that actually do not need to be crawled.
Online shops with thousands of products, product categories, filter options, search functions, etc. are particularly affected.
If you notice that new pages or updated content are taking a long time to appear in search results, it makes sense to look at your crawl budget.
Conclusion:
Don’t let SEO myths fool you! As long as your website is technically clean, has unique content, and does not contain spam, there should be no problems from Google. Then it’s all about: to inspire the user and to satisfy the search intentions. Whether a word is then in the H1 or H3 is irrelevant.
Helping Agencies & Marketers achieve their true potential with our targeted, measurable, and result-driven SEO Solutions.
Feel confident in every purchase you make with us. If for any reason we are not able to keep up to our words and you aren't satisfied, contact us within 7 days of receiving your report & we'll ensure to make it right, if not, we provide a full refund.