Reading Time: 11-13 minutes
Key Takeaways
Table of contents
Search engine optimization (SEO) is not just about helping pages rank, it is also about controlling which pages should not appear in search results. The noindex meta tag gives website owners precise control over indexation, allowing them to exclude pages that add little or no value in search.
This directive is particularly important for websites with large content libraries, eCommerce stores, or platforms that generate many low-value or duplicate pages. By applying noindex thoughtfully, site owners can keep their index lean, improve crawl efficiency, and ensure that only meaningful content is surfaced by search engines.
Google officially recognizes the robots meta tag and provides clear guidance on how it works in its documentation: Robots meta tag and X-Robots-Tag specifications .
The noindex meta tag (sometimes called meta robots noindex ) is a directive that tells search engines not to include a specific page in their search index. Unlike robots.txt, which prevents crawling, a noindex tag allows a crawler to access the page but ensures it will not appear in search results.
Here is the basic syntax:
<meta name="robots" content="noindex">
This tag can be combined with other directives such as nofollow:
<meta name="robots" content="noindex, nofollow">
“The noindex directive is one of the most powerful tools in technical SEO. It allows site owners to control exactly what appears in search results without cutting off crawlability.” — Mike Zhmudikov, SEO Director, Seologist
Not every page on your website should rank in Google. Some are essential for user experience but have no business in the search index. If indexed, these pages can create noise, dilute authority, or cause duplicate content issues.
The table below highlights the most common types of pages where applying a noindex directive is either recommended or sometimes necessary, along with the reasoning behind each case.
| Page Type | Should You Use Noindex? | Reason |
|---|---|---|
| Thank-you pages | ✅ Yes | Conversion-only, no SEO value |
| Internal search results | ✅ Yes | Thin/duplicate content |
| Login/account pages | ✅ Yes | Private, user-specific |
| Duplicate parameter URLs | ✅ Yes | Avoids index bloat |
| Category pages in eCommerce | ⚠ Sometimes | Use case dependent |
| Blog posts & landing pages | ❌ No | Valuable for SEO |
In short, the noindex tag is not meant to hide important content but to filter out pages that add little or no value to search engines. By applying it carefully, you help Google and other crawlers focus on your most strategic pages while keeping irrelevant or redundant ones out of the index.
At first glance, Noindex and Robots.txt may appear similar, but they serve different purposes in SEO and index management.
| Feature | Noindex Tag | Robots.txt Disallow |
|---|---|---|
| Blocks crawling? | ❌ No | ✅ Yes |
| Blocks indexing? | ✅ Yes | ❌ Not guaranteed |
| Works per page? | ✅ Yes | ⚠ Limited (per path) |
| Supports other directives (nofollow, etc.) | ✅ Yes | ❌ No |
| Recommended for thin/duplicate content | ✅ Yes | ❌ No |
According to
Google’s documentation
, robots.txt only controls crawling, not indexing. If you want to make sure a page does not appear in search results, a noindex directive should be used.
Robots.txt is best applied for restricting crawler access to unnecessary or sensitive resources, while noindex should be used when a page can be crawled but must stay out of the search index.
Another frequent source of confusion is the difference between noindex and canonical tags . Both are designed to deal with duplicate or overlapping content, but their purpose is not identical.
| Feature | Canonical Tag | Noindex Tag |
|---|---|---|
| Keeps a page in the index? | ✅ Yes (main version) | ❌ No |
| Passes link equity? | ✅ Yes | ❌ No |
| Best for duplicates? | ✅ Yes | ✅ Yes, when the page should be excluded entirely |
| Can be combined with others? | ✅ Yes (with hreflang, etc.) | ✅ Yes (with nofollow) |
Example canonical tag:
<link rel="canonical" href="https://www.example.com/main-page/">
So, use canonical when you want to consolidate authority between similar pages, and apply noindex when the page should be completely excluded from search visibility.
There are several ways to add a noindex directive , and the choice depends on the type of content and platform. The most common is by using the meta robots tag directly in the HTML code of a page.
To exclude a page from search results, you need to insert the following line of code into the <head> section of your HTML:
<head>
<meta name="robots" content="noindex">
</head>
Important: The tag must always be placed inside the <head> of the page, not in the <body>. If it is placed incorrectly, search engines may ignore it.
Sometimes you may want to exclude a page from indexing and prevent search engines from following links on it. In this case, you can use:
<meta name="robots" content="noindex, nofollow">
This combination is typically used for low-value pages such as login portals, checkout steps in eCommerce, or temporary test pages.
If you only want to exclude the page but still allow crawlers to follow the links on it, use:
<meta name="robots" content="noindex, follow">
This is useful for category or filter pages in online stores: the page itself should not be indexed, but the product links should still be discoverable.
For non-HTML content (like PDFs, images, or other files), you cannot insert a meta tag into the file itself. Instead, you use the X-Robots-Tag directive in the server’s HTTP response header.
Example:
X-Robots-Tag: noindex
This tells search engines that the file should not be indexed.
So, the best practice is to place a meta robots noindex directive in the <head> of the page when dealing with standard HTML content, and to use X-Robots-Tag headers for files and non-HTML resources. This way you ensure that only the right content makes it into Google’s index.
Adding a noindex tag is straightforward in most content management systems (CMS). The process differs depending on the platform, but the goal is the same — insert a meta robots directive into the page’s <head> section or configure it via built-in SEO tools or plugins.
Best use: thank-you pages, test posts, or private content.
Best use: avoid index bloat from layered navigation and duplicate categories.
Regardless of whether you use WordPress, Magento, or a website builder like Shopify or Wix, the principle is the same: you either configure noindex through an SEO plugin/module or insert the <meta> tag into the <head>. The most important step is to confirm with Google Search Console that the directive is being read correctly.
While the noindex tag is a powerful tool, it should be used with precision. Overusing it can hide valuable pages from search engines, while underusing it can lead to duplicate content and index bloat. The following best practices outline how to apply noindex effectively.
In summary, the best approach is to apply noindex selectively — only to pages that add no search value. This strategy helps maintain a clean and efficient index, reduces duplicate content, and ensures your crawl budget is focused on the pages that truly matter.
Even when applied correctly, the noindex directive can sometimes lead to confusion. Below are the most frequent mistakes and how to resolve them.
Overall, troubleshooting noindex usually comes down to understanding crawl behaviour and making sure directives are visible to Googlebot.
The noindex directive is one of the most precise tools in technical SEO. When used correctly, it helps maintain a clean, efficient index by keeping low-value or duplicate pages out of search results. This ensures that your crawl budget is spent on the content that matters most, while unnecessary pages stay hidden.
The key is balance: apply noindex to thank-you pages, faceted navigation, and thin duplicates, but avoid overusing it on potentially valuable content. Regular audits with Google Search Console can help confirm that your implementation is working as intended.
Contact a Seologist to get a tailored SEO strategy for your website.
A noindex tag is a directive placed in the HTML or HTTP header that tells search engines not to include the page in search results.
Use SEO plugins such as Yoast or RankMath. In the page settings, set the option to “No” for “Allow search engines to show this Page in search results.” This automatically generates a noindex meta tag.
Yes. A common example is:
<meta name="robots" content="noindex, nofollow">
This excludes the page from search results and also prevents crawlers from following its links.
No. It only applies to the page where it is set. However, if you use “noindex, nofollow,” internal links on that page will not pass equity.
Use the Page indexing report in Google Search Console. Pages excluded by noindex will appear under “Why pages aren’t indexed → Excluded by ‘noindex’ tag.”