Skip to content

Script to update most frequent disallowed robots.txt entries

Notifications You must be signed in to change notification settings

felmoltor/robotstxt

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Robots TXT Dictionary generator

This script requests all the robots.txt of a list of domains. Obtains the Disallow entries and order them by frequency. The resultant dictionary can be used in web fuzzing.

Usage

Usage: ./robotstxt.sh <domains file>

Output

The "output" folder will be created with the following 4 files:

  • robots.entries.txt: Contains, sorted by descending frequency the disallowed full entries of the explored domains
  • robots.sorted.files.txt: Contains, sorted by descending frequency the disallowed file names of the explored domains
  • robots.sorted.full.path.txt: Contains, sorted by descending frequency the disallowed full path entries of the explored domains
  • robots.sorted.rootfolder.txt: Contains, sorted by descending frequency the disallowed root folder entries of the explored domains

About

Script to update most frequent disallowed robots.txt entries

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages