seo(robots): disallow tag-filtered blog URLs#4247
Conversation
|
The latest updates on your projects. Learn more about Vercel for GitHub. |
PR SummaryLow Risk Overview Reviewed by Cursor Bugbot for commit eb3281b. Configure here. |
Greptile SummaryAdds Confidence Score: 5/5Safe to merge — single-line addition to an SEO config file with no logic changes. The change is minimal and correct: the wildcard pattern No files require special attention. Important Files Changed
Flowchart%%{init: {'theme': 'neutral'}}%%
flowchart TD
A[Crawler requests /robots.txt] --> B[Next.js robots function]
B --> C{URL pattern match}
C -->|matches /blog*tag=| D[Disallowed - not crawled]
C -->|matches /api/ or /workspace/ etc| D
C -->|no disallow match| E[Allowed - crawled and indexed]
D --> F[Crawler skips URL]
E --> G[Crawler indexes page]
Reviews (1): Last reviewed commit: "seo(robots): disallow tag-filtered blog ..." | Re-trigger Greptile |
Summary
Adds
Disallow: /blog*tag=to the generated robots.txt to prevent crawlers from indexing tag-filtered blog views (both/blog?tag=Xand paginated/blog?page=N&tag=X). Per SEO team request to reduce duplicate content surfaced from tag filters.Type of Change
Testing
Visit
/robots.txtafter deploy and confirm the newDisallow: /blog*tag=line is present.Checklist