Hey Folks, in this tutorial we will discuss the “web crawler” tool that will help to find important and hidden files or directories from web application. Let’s talk about the interacting and features of this tool. There is no such special feature in this tool but when we use it then it will look like both GUI and CLI 😁 . As we mentioned above it is a web crawler and web scraper tool written in Golang and may be useful for a wide range of purposes such as metadata and data extraction, data mining, reconnaissance and testing.

Lets take a example 🙂 !!

Installation

The tool is written in Golang language, so before installation we have to fulfill some dependencies of this tool and configure this one by executing the following command.

Now we install the tool from the github page and go the directory of the tool. After doing this we build the application through Go language and move it permanently after which we can easily run this tool by entering the name from anywhere.

We can see the features of this tool by executing the “help” command.

Usage

The installation was simple, but can be difficult for users to use as it has some different features which are given below.

  • Enter – Run crawler (from URL view)
  • Tab – Go to Next view
  • Ctrl+Space – Run crawler
  • Ctrl+S – Save response
  • Ctrl+Z – Quit
  • Ctrl+R – Restore to default values (from Options and Headers views)
  • Ctrl+Q – Close response save view (from Save view)

Crawling

You must have to read the above instructions because it will help to control this tool. Now here we will give the URL and click on the tab button and go to the options section and true all the options that we want to crawl.

We will not see results before entering all on the key section and after entering “all” keywords we can see that result.

Save Output

If we want to save our output or results then we need to input “Ctrl+S” through the keyboard on response section, give the file name and exit it by “Ctrl+q“.

Filters

Key sections are used to get accurate results by adding fillers. For example you can see below that after adding the PDF keyword it shows only the results that match the given keyword.

Custom Header

Some websites do not allow us to do crawling, then we use separate-2 methods and this is one of the ways in which we send different information to the server through header manipulation.

About the Author
Shubham Goyal Certified Ethical Hacker, information security analyst, penetration tester and researcher. Can be Contact on Linkedin.

Leave a Reply

Your email address will not be published. Required fields are marked *