composer require verschuur/laravel-robotstxt
Add the following to your composer.json
and then run composer install
.
{
"require": {
"verschuur/laravel-robotstxt": "^3.0"
}
}
This package supports Laravel's service provider autodiscovery so that's it. If you wish to register the package manually, add the ServiceProvider to the providers array in config/app.php
.
Verschuur\Laravel\RobotsTxt\Providers\RobotsTxtProvider::class
This package adds a /robots.txt
route to your application. Remember to remove the physical robots.txt
file from your /public
dir or else it will take precedence over Laravel's route and this package will not work.
By default, the production
environment will show
User-agent: *
Disallow:
while every other environment will show
User-agent: *
Disallow: /
This will allow the default install to allow all robots on a production environment, while disallowing robots on every other environment.
If you need custom sitemap entries, publish the configuration file
php artisan vendor:publish --provider="Verschuur\Laravel\RobotsTxt\Providers\RobotsTxtProvider"
This will copy the robots-txt.php
config file to your app's config
folder. In this file you will find the following array structure
'environments' => [
'{environment name}' => [
'paths' => [
'{robot name}' => [
'disallow' => [
''
],
'allow' => []
],
]
]
]
In which:
{environment name}
: the enviroment for which to define the paths.{robot name}
: the robot for which to define the paths.disallow
: all entries which will be used by thedisallow
directive.allow
: all entries which will be used by theallow
directive.
By default, the environment name is set to production
with a robot name of *
and a disallow entry consisting of an empty string. This will allow all bots to access all paths on the production environment.
Note: If you do not define any environments in this configuration file (i.e. an empty configuration), the default will always be to disallow all bots for all paths.
For brevity, the environment
array key will be disregarded in these examples.
Allow all paths for all robots on production, and disallow all paths for every robot in staging.
'production' => [
'paths' => [
'*' => [
'disallow' => [
''
]
]
]
],
'staging' => [
'paths' => [
'*' => [
'disallow' => [
'/'
]
]
]
]
Allow all paths for all robot bender on production, but disallow /admin
and /images
on production for robot flexo
'production' => [
'paths' => [
'bender' => [
'disallow' => [
''
]
],
'flexo' => [
'disallow' => [
'/admin',
'/images'
]
]
]
],
Besides the more standard disallow
directive, the allow
directive is also supported.
Allow a path, but disallow sub paths:
'production' => [
'paths' => [
'*' => [
'disallow' => [
'/foo/bar'
],
'allow' => [
'/foo'
]
]
]
],
When the file is rendered, the disallow
directives will always be placed before the allow
directives.
If you don't need one or the other directive, and you wish to keep the configuration file clean, you can simply remove the entire key from the entire array.
This package also allows to add sitemaps to the robots file. By default, the production environment will add a sitemap.xml entry to the file. You can remove this default entry from the sitemaps
array if you don't need it.
Because sitemaps always need to an absolute url, they are automatically wrapped using Laravel's url() helper function. The sitemap entries in the config file should be relative to the webroot.
'environments' => [
'production' => [
'sitemaps' => [
'sitemap.xml'
]
]
]
'environments' => [
'production' => [
'sitemaps' => [
'sitemap-articles.xml',
'sitemap-products.xml',
'sitemap-etcetera.xml'
]
]
]
This package is compatible with Laravel 9, 10 and 11. For a complete overview of supported Laravel and PHP versions, please refer to the 'Run test' workflow.
PHPUnit test cases are provided in /tests
. Run the tests through composer run test
or vendor/bin/phpunit --configuration phpunit.xml
.
The following reference was while creating this package: