Repository navigation

#

robots-txt

Polite, slim and concurrent web crawler.

Go
2045
4 年前

A simple and flexible web crawler that follows the robots.txt policies and crawl delays.

Go
789
4 年前

Tame the robots crawling and indexing your Nuxt site.

TypeScript
462
15 小时前

The robots.txt exclusion protocol implementation for Go language

Go
274
2 年前

A simple but powerful web crawler library for .NET

C#
251
1 年前

A set of reusable Java components that implement functionality common to any web crawler

Java
243
19 天前

Determine if a page may be crawled from robots.txt, robots meta tags and robot headers

PHP
236
3 个月前

Ultimate Website Sitemap Parser

Python
202
19 天前

Opt-Out tool to check Copyright reservations in a way that even machines can understand.

Python
194
1 年前

🤖 The largest directory for AI-ready documentation and tools implementing the proposed llms.txt standard

TypeScript
183
8 天前

Open-Source Python Based SEO Web Crawler

Python
171
2 年前

NodeJS robots.txt parser with support for wildcard (*) matching.

JavaScript
153
6 个月前

Known tags and settings suggested to opt out of having your content used for AI training.

HTML
143
10 个月前

Makes it easy to add robots.txt, sitemap and web app manifest during build to your Astro app.

TypeScript
117
1 年前

grobotstxt is a native Go port of Google's robots.txt parser and matcher library.

Go
110
3 年前

Gatsby plugin that automatically creates robots.txt for your site

JavaScript
106
1 年前

🤖 A curated list of websites that restrict access to AI Agents, AI crawlers and GPTs

Python
88
17 天前

Simple robots.txt template. Keep unwanted robots out (disallow). White lists (allow) legitimate user-agents. Useful for all websites.

86
2 个月前

ScrapeGPT is a RAG-based Telegram bot designed to scrape and analyze websites, then answer questions based on the scraped content. The bot utilizes Retrieval Augmented Generation and webscraping to return natural language answers to the user's queries.

Python
83
1 年前