robots.txt
Search engine crawler instructions.
Overview
The robots.txt file provides instructions to search engine crawlers about which pages should be indexed. It's automatically generated and accessible at /robots.txt.
How It Works
The robots.txt file:
- Allows all public pages to be crawled
- Disallows API routes and private dashboard pages
- References the sitemap location
- Uses Next.js
MetadataRoute.Robotstype
Access Points
- robots.txt - Crawler instructions for ZeroStarter
Implementation
The robots.txt is implemented in web/next/src/app/robots.ts and configures:
- User-Agent:
*(all crawlers) - Allowed: All pages under
/(public content) - Disallowed:
/api/- API endpoints/x/- Dashboard routes
- Sitemap: References
${baseUrl}/sitemap.xml
Rules
The current configuration:
- Allows all search engines (
User-Agent: *) - Permits crawling of all public pages
- Blocks API endpoints and dashboard routes
- Points crawlers to the sitemap for efficient discovery