Sign inTry Now

What does 'Internal Blocked Resource' mean in Site Audit?

Internal Blocked Resource

Description

These are internal resources (like images, CSS, or JavaScript files) that are blocked from search engine crawling. Your robots.txt file is blocking search engines from accessing important resources on your site.

How to Fix

Update your robots.txt file to allow crawling of these resources. Critical resources like CSS, JavaScript, and images should generally be accessible to search engines to ensure proper page rendering and indexing.

Detailed Analysis

Internal Blocked Resources is a common SEO issue that can significantly impact a website's performance in search engine rankings. Here's a detailed breakdown:

1. What Causes This Issue

The issue of blocked internal resources typically arises from restrictions placed on search engine crawlers through the robots.txt file. This file is used to give instructions to web crawlers about which parts of a website should not be crawled or indexed. Common causes include:

  • Misconfigured robots.txt File: Often, developers block directories containing CSS, JavaScript, or images during development and forget to update the file when the site goes live.
  • Overly Restrictive Rules: Sometimes, in an attempt to prevent sensitive or unnecessary pages from being indexed, webmasters may inadvertently block essential resources.
  • Default CMS Settings: Content Management Systems (CMS) like WordPress or Joomla may have default settings that block specific resources.
  • Incorrect Use of Directives: Misuse of directives like Disallow can lead to blocking resources that are critical for rendering pages correctly.

2. Why It's Important

Blocking internal resources can have several negative impacts:

  • Rendering Issues: Modern web pages rely heavily on CSS and JavaScript for layout and interactivity. If these resources are blocked, search engines may not be able to render the pages correctly, leading to indexing issues.
  • Affecting Mobile Usability: With Google's mobile-first indexing, blocked resources can affect how your site is viewed on mobile devices, impacting rankings.
  • Reduced Crawl Efficiency: By blocking important resources, you reduce the efficiency of search engines' ability to crawl and understand your site, potentially leading to lower rankings.
  • User Experience: Poor rendering due to blocked resources can also lead to a poor user experience, which indirectly affects SEO through increased bounce rates and lower dwell times.

3. Best Practices to Prevent It

To prevent internal resources from being blocked, consider the following best practices:

  • Review and Update the robots.txt File Regularly: Ensure that your robots.txt file is correctly configured and doesn't block essential resources.
  • Use Google Search Console: Utilize the "URL Inspection Tool" to see how Googlebot views your pages. This can help identify if any resources are being blocked inadvertently.
  • Allow Access to CSS and JS: Always ensure that CSS and JavaScript files are accessible to search engines, as they are crucial for rendering content properly.
  • Test with Different User-Agents: Use tools to simulate how different user-agents (like Googlebot) see your site to ensure no resources are unnecessarily blocked.
  • Educate Development Teams: Make sure your development teams understand the importance of search engine access to resources and the implications of blocking them.

4. Examples of Good and Bad Cases

Bad Case Example:

  • A website has a robots.txt file with the following directive:
    User-agent: *
    Disallow: /assets/
    This directive blocks the entire assets directory, which includes important CSS and JavaScript files. As a result, search engines cannot render the site's pages properly.

Good Case Example:

  • A properly configured robots.txt file looks like this:
    User-agent: *
    Allow: /assets/css/
    Allow: /assets/js/
    Disallow: /admin/
    This setup allows essential resources like CSS and JavaScript to be crawled while preventing access to non-essential or sensitive directories, such as the admin panel.

By ensuring internal resources are accessible to search engines, you can maintain a properly indexed and well-performing website. Regular audits and using tools like Google Search Console can help you stay on top of these issues.