Have you ever required to reduce Google from indexing a unique URL on your web web site and exhibiting it in their lookup motor success internet pages (SERPs)? If google serp api control web websites extended enough, a working day will probable appear when you want to know how to do this.
The 3 techniques most typically used to prevent the indexing of a URL by Google are as follows:
Applying the rel=”nofollow” attribute on all anchor elements utilised to url to the site to prevent the back links from currently being followed by the crawler.
Applying a disallow directive in the site’s robots.txt file to avoid the webpage from staying crawled and indexed.
Applying the meta robots tag with the content material=”noindex” attribute to prevent the webpage from becoming indexed.
Even though the distinctions in the 3 methods look to be refined at very first look, the success can vary considerably relying on which technique you pick out.
Employing rel=”nofollow” to avoid Google indexing
A lot of inexperienced site owners endeavor to stop Google from indexing a individual URL by applying the rel=”nofollow” attribute on HTML anchor factors. They include the attribute to each anchor factor on their web-site employed to hyperlink to that URL.
Which includes a rel=”nofollow” attribute on a website link prevents Google’s crawler from next the backlink which, in change, helps prevent them from finding, crawling, and indexing the concentrate on site. While this strategy might perform as a limited-time period alternative, it is not a practical very long-term remedy.
The flaw with this strategy is that it assumes all inbound hyperlinks to the URL will include things like a rel=”nofollow” attribute. The webmaster, nonetheless, has no way to stop other world-wide-web web-sites from linking to the URL with a adopted link. So the prospects that the URL will sooner or later get crawled and indexed using this method is quite superior.
Making use of robots.txt to stop Google indexing
One more prevalent strategy utilized to avert the indexing of a URL by Google is to use the robots.txt file. A disallow directive can be extra to the robots.txt file for the URL in concern. Google’s crawler will honor the directive which will prevent the website page from being crawled and indexed. In some conditions, having said that, the URL can still appear in the SERPs.
From time to time Google will screen a URL in their SERPs nevertheless they have never ever indexed the contents of that page. If enough net internet sites url to the URL then Google can frequently infer the matter of the webpage from the connection textual content of these inbound one-way links. As a outcome they will present the URL in the SERPs for similar lookups. While making use of a disallow directive in the robots.txt file will protect against Google from crawling and indexing a URL, it does not assurance that the URL will never look in the SERPs.
Making use of the meta robots tag to stop Google indexing
If you need to have to avoid Google from indexing a URL though also avoiding that URL from remaining exhibited in the SERPs then the most productive method is to use a meta robots tag with a information=”noindex” attribute inside of the head element of the world-wide-web site. Of class, for Google to truly see this meta robots tag they have to have to initially be equipped to find out and crawl the site, so do not block the URL with robots.txt. When Google crawls the page and discovers the meta robots noindex tag, they will flag the URL so that it will never ever be revealed in the SERPs. This is the most successful way to avert Google from indexing a URL and displaying it in their look for final results.