Best effort attempt to extract all possible domains and subdomains / FQDNs from a specified file
git clone https://github.com/intrudir/domainExtractor.git
python3 domainExtractor.py
usage: domainExtractor.py [-h] [--file INPUTFILE] [--target TARGET] [--verbose]
This script will extract domains from the file you specify and add it to a final file
optional arguments:
-h, --help show this help message and exit
--file INPUTFILE Specify the file to extract domains from
--target TARGET Specify the target top-level domain you'd like to find and extract e.g. uber.com
--verbose Enable slightly more verbose console output
python3 domainExtractor.py --file test.html --target yahoo.com
It will extract, sort and dedup all domains that are found.
Specifying 'all' as the target extracts all domains it finds (at the moment .com, .net, .org, .tv, .io)python3 domainExtractor.py --file test.html --target all
1) if you already have a final file for it it will notify you of domains you didnt have before
2) It will append them to the final file
3) It will also create a new file with the date and log them in there with the time.
This allows you to check the same target across multiple files and be notified of any new domains found!
I first use it against my Amass results, then against my Assetfinder results.
The script will sort and dedup as well as notify me of how many new, unique domains came from assetfinder's results.
It will add them to the final file as well as log just the new ones to a file with the date and time.