Generate an XML sitemap for a GitHub Pages site using GitHub Actions
.html
from URLs listed in the sitemap
for html files. GitHub Pages automatically serves a corresponding
html file if a user browses to a page with a URL with no file extension.
This new option to the generate-sitemap
action enables your sitemap to
match this behavior if you prefer the extension-less look of URLs. There
is a new action input, drop-html-extension
, to control this behavior.<meta name="robots" content="noindex">
directives, the generate-sitemap GitHub action now parses a robots.txt file, if present at the root of the website, excluding any URLs from the sitemap that match Disallow:
rules for User-agent: *
.This action generates a sitemap for a website hosted on GitHub Pages. It supports both xml and txt sitemaps. When generating an xml sitemap, it uses the last commit date of each file to generate the <lastmod>
tag in the sitemap entry. It can include html as well as pdf files in the sitemap, and has inputs to control the included file types (defaults include both html and pdf files in the sitemap). It skips html files that contain <meta name="robots" content="noindex">
. It otherwise does not currently attempt to respect a robots.txt file. The sitemap entries are sorted in a consistent order (primary sort is by depth of page in site, and URLs at same depth are then sorted alphabetically).
This action generates a sitemap for a website hosted on GitHub Pages. It supports both xml and txt sitemaps. When generating an xml sitemap, it uses the last commit date of each file to generate the <lastmod>
tag in the sitemap entry. It can include html as well as pdf files in the sitemap, and has inputs to control the included file types (defaults include both html and pdf files in the sitemap). It skips over html files that contain <meta name="robots" content="noindex">
. It otherwise does not currently attempt to respect a robots.txt file. The sitemap entries are sorted in a consistent order (primary sort is by depth of page in site, and URLs at same depth are then sorted alphabetically).
This action generates a sitemap for a website hosted on GitHub Pages. It supports both xml and txt sitemaps. When generating an xml sitemap, it uses the last commit date of each file to generate the <lastmod>
tag in the sitemap entry. It can include html as well as pdf files in the sitemap, and has inputs to control the included file types (defaults include both html and pdf files in the sitemap). It skips over html files that contain <meta name="robots" content="noindex">
. It otherwise does not currently attempt to respect a robots.txt file. The sitemap entries are sorted in a consistent order (primary sort is by depth of page in site, and URLs at same depth are then sorted alphabetically).