sel logo
Search Engine Land » SEO »
If you see “Discovered – currently not indexed” in Google Search Console, it means Google is aware of the URL, but hasn’t crawled and indexed it yet.
It doesn’t necessarily mean the page will never be processed. As their documentation says, they may come back to it later without any extra effort on your part.
But other factors could be preventing Google from crawling and indexing the page, including:
You can also use Google Search Console Inspection API to queue URLs for their coverageState
status (as well as other useful data points) en masse.
This is an obvious resolution and for the majority of cases, it will resolve the issue.
Sometimes, Google is simply slow to crawl new URLs – it happens. But other times, underlying issues are the culprit.
When you request indexing, one of two things might happen:
Both are symptoms of underlying issues.
The second happens because requesting indexing sometimes gives your URL a temporary “freshness boost” which can take the URL above the requisite quality threshold and, in turn, lead to temporary indexing.
Get the daily newsletter search marketers rely on.
Processing…Please wait.
See terms.
This is where vocabulary can get confusing. I’ve been asked, “How can Google determine the page quality if it hasn’t been crawled yet?”
This is a good question, and the answer is that it can’t.
Google is making an assumption about the page’s quality based on other pages on the domain. Their classifications are likewise based on URL patterns and website architecture.
As a result, moving these pages from “awareness” to the crawl queue can be de-prioritized based on the lack of quality they have found on similar pages.
It’s possible that pages on similar URL patterns or those located in similar areas of the site architecture have a low-value proposition compared to other pieces of content targeting the same user intents and keywords.
Possible causes include:
Working on improving the content quality within the site cluster and the specific pages can have a positive impact on reigniting Google’s interest in crawling your content with greater purpose.
You can also noindex other pages on the website that you acknowledge aren’t of the highest quality to improve the ratio of good-quality pages to bad-quality pages on the site.
Crawl budget is an often misunderstood mechanism in SEO.
The majority of websites don’t need to worry about this. In fact, Google’s Gary Illyes has gone on the record claiming that probably 90% of websites don’t need to think about crawl budget. It is often regarded as a problem for enterprise websites.
Crawl efficiency, on the other hand, can affect websites of all sizes. Overlooked, it can lead to issues on how Google crawls and processes the website.
To illustrate, if your website:
…then you might be having duplication issues that impact Google’s assumptions on crawl priority based on wider site assumptions.
You might be zapping Google’s crawl budget with unnecessary URLs and requests. Given that Googlebot crawls websites in portions, this can lead to Google’s resources not stretching far enough to discover all newly published URLs as fast as you would like.
You want to crawl your website regularly, and ensure that:
If your website utilizes parameters, such as ecommerce product filters, you can curb the crawling of these URI paths by disallowing them in the robots.txt file.
Your server can also be important in how Google allocates the budget to crawl your website.
If your server is overloaded and responding too slowly, crawling issues may arise. In this case, Googlebot won’t be able to access the page resulting in some of your content not getting crawled.
Consequently, Google will try to come back later to index the website, but it will no doubt cause a delay in the whole process.
When you have a website, it’s important to have internal links from one page to another.
Google usually pays less attention to URLs that don’t have any or enough internal links – and may even exclude them from its index.
You can check the number of internal links to pages through crawlers like Screaming Frog and Sitebulb.
Having an organized and logical website structure with internal links is the best way to go when it comes to optimizing your website.
But if you have trouble with this, one way to make sure all of your internal pages are connected is to “hack” into the crawl depth using HTML sitemaps.
These are designed for users, not machines. Although they may be seen as relics now, they can still be useful.
Additionally, if your website has many URLs, it’s wise to split them up among multiple pages. You don’t want them all linked from a single page.
Internal links also need to use the <a> tag for internal links instead of relying on JavaScript functions such as onClick()
.
If you’re utilizing a Jamstack or JavaScript framework, investigate how it or any related libraries handle internal links. These must be presented as <a> tags.
Opinions expressed in this article are those of the guest author and not necessarily Search Engine Land. Staff authors are listed here.
Related stories
New on Search Engine Land
About the author
Related topics
Get the daily newsletter search marketers rely on.
See terms.
Learn actionable search marketing tactics that can help you drive more traffic, leads, and revenue.
March 15-16, 2023: SMX Munich
Online Aug. 16-17: SMX Master Classes
Online Nov. 14-15: SMX Next
Start training now: SMX Advanced
Discover time-saving technologies and actionable tactics that can help you overcome crucial marketing challenges.
April 15-17, 2020: San Jose
Become a Data Champion By Mastering Marketing Analytics Automation
Why Consent and Compliance Matter With Programmatic Ads
From Zero to Hero: Apply Pillar-Based Marketing to Your SEO and Digital Campaigns to Drive Revenue
Enterprise Marketing Attribution and Performance Management Platforms: A Marketer’s Guide
Enterprise Account-Based Marketing Platforms: A Marketer’s Guide
Enterprise Marketing Work Management Platforms: A Marketer’s Guide
The Content Marketer’s Funnel
Meet your new AI-powered marketing assistant!
Get the must-read newsletter for search marketers.
Topics
Our events
About
Follow us
© 2023 Third Door Media, Inc. All rights reserved.
Third Door Media, Inc. is a publisher and marketing solutions provider incorporated in Delaware, USA, with an address 88 Schoolhouse Road, PO Box 3103, Edgartown, MA 02539. Third Door Media operates business-to-business media properties and produces events. It is the publisher of Search Engine Land the leading Search Engine Optimization digital publication.