spatie/robots-txt
Generate and serve a correct robots.txt in Laravel with an expressive API. Add user-agents, allow/disallow rules, sitemaps and host directives, then publish it via a route or controller—perfect for managing crawler access per environment.
|
Package
|
Description
|
Stars
|
Likes
|
Forks
|
Downloads
|
Issues
|
Score
|
Opportunity
|
License
|
Last Release
|
|
|---|---|---|---|---|---|---|---|---|---|---|---|
| spatie/crawler | Fast, concurrent web crawler for PHP. Crawl sites, collect internal URLs with depth limits, and hook into crawl events. Can execute JavaScript via Chrome/Puppeteer for rendered pages. Includes fakes for testing crawl logic without real HTTP requests. | 2,803 | 2,801 | 368 | 627K | 0 | 55.7 | 38.2 | MIT | 2 weeks ago | |
| spatie/one-package-to-rule-them-all | Pulls in all Spatie packages at once so you don’t have to require them individually. Install via Composer, optionally publish the config, and get the full Spatie “magic” in one go for your Laravel app. | 60 | 60 | 6 | 0 | 0 | 20.8 | — | MIT | 5 years ago |
How can I help you explore Laravel packages today?