Skip to main content
Settings
Search
Appearance
Theme Mode
About
Jekyll v3.10.0
Environment Production
Last Build
2026-05-06 02:34 UTC
Current Environment Production
Build Time May 06, 02:34
Jekyll v3.10.0
Build env (JEKYLL_ENV) production
Quick Links
Page Location
Page Info
Layout article
Collection posts
Path _posts/data-analytics/2023-12-04-robots-txt-jekyll.md
URL /posts/robots-txt-jekyll/
Date 2023-12-04
Theme Skin
SVG Backgrounds
Layer Opacity
0.6
0.04
0.08

robots dot txt

By bamr87

Learn how to properly configure robots.txt files for Jekyll websites to control search engine crawlers and optimize SEO

Estimated reading time: 4 minutes

Table of Contents

robots dot txt

A robots.txt file is used to instruct web robots (typically search engine robots) which pages on your website to crawl and which not to. Here’s an example of a robots.txt file that you might use for a Jekyll site:

User-agent: *
Disallow: /secret/
Disallow: /private/
Disallow: /tmp/
Sitemap: https://it-journey.dev/sitemap.xml

In this example:

  • User-agent: * means that the instructions apply to all web robots.
  • Disallow: /secret/ tells robots not to crawl pages under the /secret/ directory.
  • Disallow: /private/ and Disallow: /tmp/ do the same for these directories.
  • Sitemap: https://www.yoursite.com/sitemap.xml provides the location of your site’s sitemap, which is helpful for search engines to find and index your content.

Remember to replace https://www.yoursite.com/sitemap.xml with the actual URL of your sitemap. Also, the Disallow entries should be adjusted based on the specific directories or pages you want to keep private or don’t want to be indexed by search engines.