ParamSpider Save

Mining URLs from dark corners of Web Archives for bug hunting/fuzzing/further probing

Project README

paramspider

Mining URLs from dark corners of Web Archives for bug hunting/fuzzing/further probing

📖 About🏗️ Installation⛏️ Usage🚀 Examples🤝 Contributing

paramspider

About

paramspider allows you to fetch URLs related to any domain or a list of domains from Wayback achives. It filters out "boring" URLs, allowing you to focus on the ones that matter the most.

Installation

To install paramspider, follow these steps:

git clone https://github.com/devanshbatham/paramspider
cd paramspider
pip install .

Usage

To use paramspider, follow these steps:

paramspider -d example.com

Examples

Here are a few examples of how to use paramspider:

  • Discover URLs for a single domain:

    paramspider -d example.com
    
  • Discover URLs for multiple domains from a file:

    paramspider -l domains.txt
    
  • Stream URLs on the termial:

    paramspider -d example.com -s
    
  • Set up web request proxy:

    paramspider -d example.com --proxy '127.0.0.1:7890'
    
  • Adding a placeholder for URL parameter values (default: "FUZZ"):

     paramspider -d example.com -p '"><h1>reflection</h1>'
    

Contributing

Contributions are welcome! If you'd like to contribute to paramspider, please follow these steps:

  1. Fork the repository.
  2. Create a new branch.
  3. Make your changes and commit them.
  4. Submit a pull request.

Star History

Star History Chart

Open Source Agenda is not affiliated with "ParamSpider" Project. README Source: devanshbatham/ParamSpider
Stars
2,185
Open Issues
17
Last Commit
4 days ago
License

Open Source Agenda Badge

Open Source Agenda Rating