How to Fix website Index error Easy ?
SEO Help and Tips
How to Fix Index error Easy ?
Detect where the noindex directives are coming from. They can be added through various methods, such as meta tags, robots.txt file.
Remove meta tags
with noindex and nosnippet:
If the noindex directives are present in the meta tags of
your web pages, locate the meta
tags in the HTML
code of each page and remove the "noindex" value. Look for the
following
meta tag. Simply delete or modify this meta tag to remove
the noindex directive. Make sure
to repeat this process for all relevant pages.
{ <meta name="robots"
content="nofollow"> }: This Tag instructs search engine
robots not to
follow any links on the web page. This means that search
engine crawlers should not crawl
or index the pages that are linked from the current web
page. The "nofollow" directive is
commonly used for links that are intended for user
interaction (such as comments or
advertisements) and should not be given weight or
influence in search engine rankings.
{ <meta name="robots"
content="noindex"> }:This tag tells search engine robots not
to index the
current web page. This means that the page will not be
included in search engine results.
It is typically used for pages that are not meant to be
publicly accessible or are duplicate
content. Search
engine crawlers will still follow and crawl the links on the page, but the
page itself will not appear in search engine listings.
{ <meta name="googlebot"
content="noindex"> }: This tag tells the Googlebot crawler
not to
index the page in Google's search results. It instructs
the crawler to ignore the page and
not include it in
the search engine index.
{ <meta name="googlebot-news"
content="noindex"> }: This tag specifically targets the
Google
News crawler. It indicates that the page should not be
indexed by Google News
preventing it from appearing in the news search results.
{ <meta name="googlebot"
content="noindex"> }: This is another instance of the first
tag
mentioned. It
reiterates the instruction to the Googlebot crawler not to index the page.
{ <meta name="googlebot-news"
content="nosnippet"> }: This tag is used to prevent Google
News from displaying a snippet of the page's content in
the search results. It means that
when the page appears in the news search results, Google
won't show a preview snippet
of the content.
{ <meta name="robots"
content="max-snippet:0"> }: instructs search engine
robots to display
only a minimal or no snippet in search engine results.
The "max-snippet" directive allows
webmasters to control the length of the description or snippet shown in search engine listings.
Setting the value to "0" indicates that the
webmaster prefers search engines to not display
any snippet for the page. This can be useful for pages
that contain sensitive or confidential
information that should not be exposed in search results.
{ <meta content='nositelinkssearchbox' name='google'> }: nositelinkssearchbox is The value indicates that the webpage does not want to display a
sitelinks search box in Google's search results.
Sitelinks search box is a feature that allows
users to search within a specific website directly from
the search results page, providing a
more convenient and focused search experience.
By using nositelinkssearchbox, webmasters are indicating
that they do not want Google to
display this search box feature for their webpage. This
might be because the website's
own search functionality is already robust enough, or the
webmaster prefers a different
user experience when it comes to search functionality on
their website.
Check robots.txt
file:
Open the robots.txt file of your website and review its
content. Look for any directives that
disallow search engines from indexing certain pages or
sections. Remove or modify any
disallow rules that prevent indexing.
For example, if you find a line like this:
“Disallow:
/example-page/ “
to
“Allow:
/example-page/”
Web page all contents allowed crawl to Index
The default robots.txt file allows search engine crawle
to access all pages and posts on the
website. Here is the default robots.txt
“User-agent:
Mediapartners-Google
Disallow:
User-agent: *
Disallow: /search
Allow: /
Sitemap:
[Your-Blog-URL]/sitemap.xml”
The robots.txt provided allows specific user-agents to
access different types of content on
the website. Here is a breakdown of the directives in the
robots.txt file:
Which Robts.txt should use or not ?
User-agent: *
(This directive
applies to all other user-agents crawlers)
Allow: /
(This allows crawling of all other pages on the website)
Disallow: /search
(This disallows crawling of
pages under the /search directory)
User-agent: Mediapartners-Google (used for
Google AdSense)
User-agent: Googlebot
(used for regular web crawling)
Disallow: /nogooglebot/"
(used for which parts of a website
they should not crawl)
User-agent: Adsbot-Google
(used for AdWords campaigns)
User-agent: Googlebot-News
(used for crawling news content)
User-agent: Googlebot-Image
(used for crawling images)
User-agent: Googlebot-Video
(used for crawling video content)
User-agent: Googlebot-Mobile (used for crawling mobile content)
For all of the specified user-agents, no specific
disallow rules are provided, which means
they are allowed to crawl all content on the website.
This robots.txt file allows specific Google user-agents
to access all content on the website,
while other user-agents are allowed to access all content
except for pages under the
search directory.
{ <meta name="robots" content="nofollow"> }:
This Tag instructs search engine robots not to follow any links on the web page. This means that search engine crawlers should not crawl or index the pages that are linked from the current web page. The "nofollow" directive is commonly used for links that are intended for user interaction (such as comments or advertisements) and should not be given weight or influence in search engine rankings.
{ <meta name="robots" content="noindex"> }:This tag tells search engine robots not to index the current web page. This means that the page will not be included in search engine results. It is typically used for pages that are not meant to be publicly accessible or are duplicate content. Search engine crawlers will still follow and crawl the links on the page, but the page itself will not appear in search engine listings.
{ <meta name="googlebot" content="noindex"> }: This tag tells the Googlebot crawler not to index the page in Google's search results. It instructs the crawler to ignore the page and not include it in the search engine index.
{ <meta name="googlebot-news" content="noindex"> }:This tag specifically targets the Google News crawler. It indicates that the page should not be indexed by Google News preventing it from appearing in the news search results.
{ <meta name="googlebot" content="noindex"> }: This is another instance of the first tag mentioned. It reiterates the instruction to the Googlebot crawler not to index the page.
{ <meta name="googlebot-news" content="nosnippet"> }: This tag is used to prevent Google News from displaying a snippet of the page's content in the search results. It means that when the page appears in the news search results, Google won't show a preview snippet of the content.
Check robots.txt file: Open the robots.txt file of your website and review its content. Look for any directives that disallow search engines from indexing certain pages or sections.
Remove or modify anydisallow rules that prevent indexing. For example, if you find a line like this:
“Disallow: /example-page/ “
to
“Allow: /example-page/”
Remove Canonical:
Here's an example of how the canonical meta tag would look for a webpage with the
URL "https://www.example.com/blog/my-article" and "https://www.example.com/original-
page-url"
{ <link rel="canonical" href="https://www.example.com/original-page-url"> }
{ <link rel="canonical" href="https://www.example.com/blog/my-article"> }
Be patient and monitor the indexing status using Google Search Console or other similar tools.
Comments
Post a Comment
Thanks for your Comments.