> cat /dev/github | grep security-tools
discovered 30 Mar 2026

ParamSpider

Python โ˜… 3030 via github-topic
โ†’ View on GitHub

AI Summary: ParamSpider is a tool designed for fetching and filtering URLs from Wayback Machine archives, specifically targeting domains for bug hunting, fuzzing, and further probing. Its notable features include the ability to ignore less relevant URLs, support for multiple domains, proxy configuration, and the option to introduce placeholders for parameter values, enhancing the efficiency of security assessments.


README

paramspider

Mining URLs from dark corners of Web Archives for bug hunting/fuzzing/further probing

๐Ÿ“– About โ€ข ๐Ÿ—๏ธ Installation โ€ข โ›๏ธ Usage โ€ข ๐Ÿš€ Examples โ€ข ๐Ÿค Contributing โ€ข

paramspider

About

paramspider allows you to fetch URLs related to any domain or a list of domains from Wayback achives. It filters out “boring” URLs, allowing you to focus on the ones that matter the most.

Installation

To install paramspider, follow these steps:

git clone https://github.com/devanshbatham/paramspider
cd paramspider
pip install .

Usage

To use paramspider, follow these steps:

paramspider -d example.com

Examples

Here are a few examples of how to use paramspider:

  • Discover URLs for a single domain:

    paramspider -d example.com
    
  • Discover URLs for multiple domains from a file:

    paramspider -l domains.txt
    
  • Stream URLs on the termial:

    paramspider -d example.com -s
    
  • Set up web request proxy:

    paramspider -d example.com --proxy '127.0.0.1:7890'
    
  • Adding a placeholder for URL parameter values (default: “FUZZ”):

     paramspider -d example.com -p '"><h1>reflection</h1>'
    

Star History

Star History Chart