Getting Started with Burpsuite & Running a basic Web-Spider
Burpsuite is a collection of tools bundled into a single suite made for Web Application Security or Penetration testing. Its a java executable and hence its cross platform. Kali Linux comes with Buprsuite free edition installed. There is also a professional version available. The main features of burpsuite is that it can function as an intercepting proxy. Burpsuite intercepts the traffic between a web browser and the web server.
- Application Aware Spider : Used for spidering/crawling a given scope of pages.
- Scanner : Automatically scans for vulnerabilities just like any other automated scanners
- Intruder : Used to perform attacks & bruteforces on pages in a highly customize-able manner.
- Repeater : Used for manipulating and resending individual requests.
- Sequencer : Used mainly for testing/fuzzing session tokens.
- Extensibility, allowing you to easily write your own plugins, to perform complex and highly customized tasks within Burp.
- Comparer & Decoder used for misc purposes that might come along the way when you conduct a Web Security test
Spidering a Website
A web crawler is a bot program which systematically browses the pages of a website for the purpose of indexing. Precisely a web crawler maps the structure of a website by browsing all its inner pages. The crawler is also reffered to as spider or automatic indexer.
Burpsuite has got its own spider called the burpspider. The burp spider is a program which crawls into all the pages of a target specified in the scope. Before starting the burp spider, burpsuite has to to be configured to intercept the HTTP traffic.
Interface & Options
Like any other GUI/Windows tool, burpsuite contains a standard menu bar, 2 rows of tabs & different set of panels as seen below.
The above figure shows the options & details about the target. In the above figure there are mainly 4 sections. They are described against the corresponding numbers as follows:
- Tool & Options selector Tabs – Select between Various tools & settings of burpsuite
- Sitemap View – Displays the sitemap once spider has started
- Requests Queue – Displays the requests being made
- Request/Response Details – The HTTP requests made & the responses from the servers.
Lab 1 : Spidering a website
Spidering is a major part of recon while performing Web security tests. It helps the pentester to identify the scope & archetecture of the web-application.As described earlier, burpsuite has it’s own spider called the burp spider which can crawl into a website.
Scenario: Attacker – Kali Linux VM, IP = 192.168.0.105
Target – OWASP Broken Web Application VM, IP = 192.168.0.160
Step 1 : Setup Proxy.
First start burpsuite and check details under proxy tab in Options sub-tab. Ensure IP is localhost IP & port is 8080.
Also ensure that Intercept is ON in the Intercept Sub-Tab
Then on IceWeasel/Firefox, Goto Options > Preferences > Network > Connection Settings.
Choose Manual Proxy Configuration
If you want, you can try installing proxy add-ons. Here is one such.
Install the proxy selector from addons page and goto preferences
Goto Manage Proxies & add a new proxy filling out the relevant information. It’s simple.
Click the Proxy Selector button at the Top right & select the Proxy you just created.
Step 2 : Getting Content into Burp
After you have setup the proxy, goto the target normally by entering the URL in the address bar. You can notice that the page will not be loading up. This is because burpsuite is intercepting the connection.
Meanwhile in burpsuite, you can see the request details. Click forward to forward the connection. Then you can see that the page has loaded up in the browser.
Comming back to burpsuite, you can see that all sections are populated.
Step 3 : Scope Selection & Starting Spider
Now narrow down the target as you want. Here the target/mutillidae is selected. Right click the mutillidae from the sitemap & select Spider from Here option
After the spider starts, You get a prompt as shown in the following figure. It’s a login form. If you know the details, fill in as needed & thus the spider wil be able to crawl from the inside also. You can skip this step by pressing the Ignore Form button.
Step 4 : Manipulating Details
Now you can see as the spider runs, the tree inside of the mutillidae branch gets populated. Also the requests made are shown in the queue and the details are shown in the Request tab.
Move on to different Tabs and see all the underlying information.
Finally check if spider is finished by viewing the Spider tab.
These are the very basics & starting point of a web security test. Spidering is an important part of the recon during the test and by clearly executing this, we can understand about the architecture of the target site. In upcomming tutorials, we will extend this to other tools in the Burpsuite set of tools.
Source :– http://kalilinuxtutorials.com/burpsuite/
Like Our Page on Facebook :– Grey Hat Hackers
NOTE: This is for educational purpose only we are not responsible for any type of inconvenience caused by reader.