How to Fix website Index error Easy ?

SEO Help and Tips

How to Fix Index error Easy ?

Detect where the noindex directives are coming from. They can be added through various methods, such as meta tags, robots.txt file.

Remove meta tags with noindex and nosnippet: 

If the noindex directives are present in the meta tags of your web pages, locate the meta

 tags in the HTML code of each page and remove the "noindex" value. Look for the following

meta tag. Simply delete or modify this meta tag to remove the noindex directive. Make sure

to repeat this process for all relevant pages.

{ <meta name="robots" content="nofollow"> }: This Tag instructs search engine robots not to

follow any links on the web page. This means that search engine crawlers should not crawl

or index the pages that are linked from the current web page. The "nofollow" directive is

commonly used for links that are intended for user interaction (such as comments or

advertisements) and should not be given weight or influence in search engine rankings.

 

{ <meta name="robots" content="noindex"> }:This tag tells search engine robots not to index the

current web page. This means that the page will not be included in search engine results.

It is typically used for pages that are not meant to be publicly accessible or are duplicate

 content. Search engine crawlers will still follow and crawl the links on the page, but the

page itself will not appear in search engine listings.

 

{ <meta name="googlebot" content="noindex"> }: This tag tells the Googlebot crawler not to

index the page in Google's search results. It instructs the crawler to ignore the page and

 not include it in the search engine index.

 

{ <meta name="googlebot-news" content="noindex"> }: This tag specifically targets the Google

News crawler. It indicates that the page should not be indexed by Google News

preventing it from appearing in the news search results.

 

{ <meta name="googlebot" content="noindex"> }: This is another instance of the first tag

 mentioned. It reiterates the instruction to the Googlebot crawler not to index the page.

 

{ <meta name="googlebot-news" content="nosnippet"> }: This tag is used to prevent Google

News from displaying a snippet of the page's content in the search results. It means that

when the page appears in the news search results, Google won't show a preview snippet

of the content.

 

{ <meta name="robots" content="max-snippet:0"> }: instructs search engine robots to display

only a minimal or no snippet in search engine results. The "max-snippet" directive allows

webmasters to control the length of the description or snippet shown in search engine listings.

Setting the value to "0" indicates that the webmaster prefers search engines to not display

any snippet for the page. This can be useful for pages that contain sensitive or confidential

information that should not be exposed in search results.

 

{ <meta content='nositelinkssearchbox' name='google'> }:  nositelinkssearchbox is The value indicates that the webpage does not want to display a

sitelinks search box in Google's search results. Sitelinks search box is a feature that allows

users to search within a specific website directly from the search results page, providing a

more convenient and focused search experience.

By using nositelinkssearchbox, webmasters are indicating that they do not want Google to

display this search box feature for their webpage. This might be because the website's

own search functionality is already robust enough, or the webmaster prefers a different

user experience when it comes to search functionality on their website.

Check robots.txt file: 

Open the robots.txt file of your website and review its content. Look for any directives that

disallow search engines from indexing certain pages or sections. Remove or modify any

disallow rules that prevent indexing.

 

For example, if you find a line like this:

“Disallow: /example-page/ “

to 

“Allow: /example-page/”

Web page all contents allowed crawl to Index

The default robots.txt file allows search engine crawle to access all pages and posts on the

website. Here is the default robots.txt

“User-agent: Mediapartners-Google

Disallow:

User-agent: *

Disallow: /search

Allow: /

Sitemap: [Your-Blog-URL]/sitemap.xml”

 

The robots.txt provided allows specific user-agents to access different types of content on

the website. Here is a breakdown of the directives in the robots.txt file:

Which Robts.txt should use or not ?

User-agent: *                              (This directive applies to all other user-agents crawlers)

Allow: /                                        (This allows crawling of all other pages on the website)

Disallow: /search                        (This disallows crawling of pages under the /search directory)

User-agent: Mediapartners-Google   (used for Google AdSense)

User-agent: Googlebot                       (used for regular web crawling)

Disallow: /nogooglebot/"                    (used for which parts of a website they should not crawl)

User-agent: Adsbot-Google               (used for AdWords campaigns)

User-agent: Googlebot-News            (used for crawling news content)

User-agent: Googlebot-Image           (used for crawling images)

User-agent: Googlebot-Video            (used for crawling video content)

User-agent: Googlebot-Mobile          (used for crawling mobile content)


For all of the specified user-agents, no specific disallow rules are provided, which means 

they are allowed to crawl all content on the website.

This robots.txt file allows specific Google user-agents to access all content on the website,

while other user-agents are allowed to access all content except for pages under the 

search directory.

 

{ <meta name="robots" content="nofollow"> }: 

This Tag instructs search engine robots not to follow any links on the web page. This means that search engine crawlers should not crawl or index the pages that are linked from the current web page. The "nofollow" directive is commonly used for links that are intended for user interaction (such as comments or advertisements) and should not be given weight or influence in search engine rankings.

{ <meta name="robots" content="noindex"> }:This tag tells search engine robots not to index the current web page. This means that the page will not be included in search engine results. It is typically used for pages that are not meant to be publicly accessible or are duplicate content. Search engine crawlers will still follow and crawl the links on the page, but the page itself will not appear in search engine listings.

{ <meta name="googlebot" content="noindex"> }: This tag tells the Googlebot crawler not to index the page in Google's search results. It instructs the crawler to ignore the page and not include it in the search engine index.

{ <meta name="googlebot-news" content="noindex"> }:This tag specifically targets the Google News crawler. It indicates that the page should not be indexed by Google News preventing it from appearing in the news search results.

{ <meta name="googlebot" content="noindex"> }: This is another instance of the first tag mentioned. It reiterates the instruction to the Googlebot crawler not to index the page.

{ <meta name="googlebot-news" content="nosnippet"> }: This tag is used to prevent Google News from displaying a snippet of the page's content in the search results. It means that when the page appears in the news search results, Google won't show a preview snippet of the content.

Check robots.txt file: Open the robots.txt file of your website and review its content. Look for any directives that disallow search engines from indexing certain pages or sections. 

Remove or modify anydisallow rules that prevent indexing. For example, if you find a line like this:

“Disallow: /example-page/ “

to 

“Allow: /example-page/”

Remove Canonical:

Here's an example of how the canonical meta tag would look for a webpage with the 

URL "https://www.example.com/blog/my-article" and  "https://www.example.com/original-

page-url"

<link rel="canonical" href="https://www.example.com/original-page-url"> }

<link rel="canonical" href="https://www.example.com/blog/my-article"> }

Be patient and monitor the indexing status using Google Search Console or other similar tools.

Comments

Popular posts from this blog

Office Tool_SPSS v23 + Serial key

How to Fix FATAL error Failed to parse input Document ?

How to Reduce Lazy Load Resources

Popular posts from this blog

Office Tool_SPSS v23 + Serial key

How to Fix FATAL error Failed to parse input Document ?

How to Reduce Lazy Load Resources