Directories/URLs gathering
getallurls/gau, hawkrawler, gobuster, wfuzz, dirb, dirbuster, ffuf
Common web directories to recon:
Robots.txt
Sitemap.xml
Burp suite target -> site map
The Burp Suite Site map feature found under the Target tab provides an overview of the directories found by Burp, that is gathered as the web app is explored, and with additional crawling.
getallurls/gau
The package getallurls
installed directly in Kali Linux with apt install
seems to have some issue. To use the gau
binary that can be installed from the installation step listed in the Github link below.
Basic commands:
hakrawler
ffuf
Ffuf is a fast web fuzzer for directory discovery. The term fuzzing refers to the act of sending random data to applications (URLs in this case), to discover content that would not have been discovered otherwise.
Basic command:
Flags
-w:
Path to wordlist
-u
: HTTP/HTTPS endpoint URL to fuzz
The FUZZ
keyword in the URL supplied to the -u
flag will be replaced by each word given in the wordist.
Example
Suppose there is a target at the HTTP address http://88.88.88.88 to be fuzzed: with the wordlist in the ~/wordlists/common.txt directory containing common directory values.
The output shows that the directories: assets and robots.txt, returned a valid status code, indicating that there are contents present.
dirb
"DIRB is a Web Content Scanner. It looks for existing (and/or hidden) Web Objects. It basically works by launching a dictionary based attack against a web server and analyzing the responses."
Basic command:
Example
Scan the URL http://88.88.88.88/ with the wordlist provided.
gobuster
"Gobuster is a tool used to brute-force: URIs (directories and files) in web sites, DNS subdomains (with wildcard support), Virtual Host names on target web servers, ..."
Basic command:
Flags
dir
: Uses directory/file enumeration mode
--url:
HTTP/HTTPS endpoint URL to brute-force
-w
: Path to wordlist
Example
To brute-force the HTTP URL http://88.88.88.88/ with the wordlist ~/wordlists/common.txt
The output shows that the directories: assets and robots.txt, returned a valid status code, indicating that there are contents present.
Last updated