Post by account_disabled on Dec 5, 2023 2:21:24 GMT -8
Here you can see the number of pages Google crawls every day. Crawl Budget Google Search Console In the example we see that the initial crawl budget is around 300 pages per day, while the current one is around 100 pages. Determine Crawl Budget with Server Logs Checking server logs is very useful to see how often Google crawlers crawl our website. Crawl Budget Analysis Log Server Furthermore, it is interesting to compare these statistics with those reported in Search Console as it is always better to rely on multiple sources. How to optimize your Crawl Budget for SEO Crawl budget optimization is about ensuring that there is no waste in crawling a site's resources.
We must therefore be able to correct all those problems Telegram Number that can lead to a waste of crawl budget. After optimizing thousands of sites, I realized that most have the following types of problems: Accessible URL parameters Duplicate content Low quality content Broken links and Redirects Incorrect URLs in XML sitemaps Pages with long loading times or giving timeouts Large number of non-indexable pages Incorrect internal link structure Let's look at each individual problem in detail. Accessible URL parameters An example of a URL with parameter is the following: categoria/page.php?colore=nero. Here the parameter is used to store a visitor's selection in a product filter. In most cases, URLs with parameters should not be accessible to search engines , as they can generate a virtually infinite amount of URLs. Parametric URLs are commonly used when implementing product filters on e-commerce sites. It's okay to use them, but make sure they are not accessible to crawlers. To make them inaccessible to search engines: Use the robots.txt file to tell search engines not to crawl these URLs.
Alternatively, use the URL parameter management settings in Search Console and Bing Webmaster Tools to tell Google and Bing which pages should not be crawled; Add rel="nofollow" to filter links. Duplicate content By the term “duplicate content” we refer to pages that are identical or very similar to each other. Examples of duplicate content include: pages with content copied from other sites, internal search results pages, and tag pages. Your goal is to ensure that the search engine does not waste time crawling pages with duplicate content , which could negatively impact the allocated budget. In this sense, it becomes essential to ensure that there is no duplicate content on your site, or at least that the presence of such content is reduced to a minimum. How to do? Implement 301 redirects for all domain variants internal search results pages inaccessible to Google using the robots.txt file; Disable pages dedicated to images (for example: the infamous attachment pages in WordPress); Use category pages and tags correctly. Low quality content Pages with sparse or inconsistent content that add no value to users are not attractive to search engines and you should avoid publishing such pages on your site .