summaryrefslogtreecommitdiffstats
path: root/docs/content/en/templates/robots.md
diff options
context:
space:
mode:
Diffstat (limited to 'docs/content/en/templates/robots.md')
-rw-r--r--docs/content/en/templates/robots.md66
1 files changed, 66 insertions, 0 deletions
diff --git a/docs/content/en/templates/robots.md b/docs/content/en/templates/robots.md
new file mode 100644
index 000000000..52a77314b
--- /dev/null
+++ b/docs/content/en/templates/robots.md
@@ -0,0 +1,66 @@
+---
+title: Robots.txt File
+linktitle: Robots.txt
+description: Hugo can generate a customized robots.txt in the same way as any other template.
+date: 2017-02-01
+publishdate: 2017-02-01
+lastmod: 2017-02-01
+categories: [templates]
+keywords: [robots,search engines]
+menu:
+ docs:
+ parent: "templates"
+ weight: 165
+weight: 165
+sections_weight: 165
+draft: false
+aliases: [/extras/robots-txt/]
+toc: false
+---
+
+To generate a robots.txt file from a template, change the [site configuration][config]:
+
+{{< code-toggle file="config">}}
+enableRobotsTXT = true
+{{< /code-toggle >}}
+
+By default, Hugo generates robots.txt using an [internal template][internal].
+
+```text
+User-agent: *
+```
+
+Search engines that honor the Robots Exclusion Protocol will interpret this as permission to crawl everything on the site.
+
+## Robots.txt Template Lookup Order
+
+You may overwrite the internal template with a custom template. Hugo selects the template using this lookup order:
+
+1. `/layouts/robots.txt`
+2. `/themes/<THEME>/layouts/robots.txt`
+
+## Robots.txt Template Example
+
+{{< code file="layouts/robots.txt" download="robots.txt" >}}
+User-agent: *
+{{ range .Pages }}
+Disallow: {{ .RelPermalink }}
+{{ end }}
+{{< /code >}}
+
+This template creates a robots.txt file with a `Disallow` directive for each page on the site. Search engines that honor the Robots Exclusion Protocol will not crawl any page on the site.
+
+{{% note %}}
+To create a robots.txt file without using a template:
+
+1. Set `enableRobotsTXT` to `false` in the [site configuration][config].
+2. Create a robots.txt file in the `static` directory.
+
+Remember that Hugo copies everything in the [static directory][static] to the root of `publishDir` (typically `public`) when you build your site.
+
+[config]: /getting-started/configuration/
+[static]: /getting-started/directory-structure/
+{{% /note %}}
+
+[config]: /getting-started/configuration/
+[internal]: https://github.com/gohugoio/hugo/blob/master/tpl/tplimpl/embedded/templates/_default/robots.txt