First, I must create a dictionary although it will not be used in this test. One interesting option that I have chosen is -I, in order to only follow those URLs which match the string associated with the parameter.
javi@ubuntu-server:~/skipfish-1.86b$ cp -a dictionaries/complete.wl dictionary.wl javi@ubuntu-server:~/skipfish-1.86b$ ./skipfish -W /dev/null -I http://192.168.122.104/mediawiki -o mediawiki_dir http://192.168.122.104/mediawiki
If you do not set this option and skipfish figures out more sites, it will scan them as well. In case you want to shut out a specific URL, you must establish it by means of the -X parameter.
During the crawling, skipfish shows information in real time about its analysis.
skipfish version 1.86b by <email@example.com> - 192.168.122.104 - Scan statistics: Scan time : 0:59:38.542 HTTP requests : 34669 (10.4/s), 100769 kB in, 12487 kB out (31.6 kB/s) Compression : 77967 kB in, 255451 kB out (53.2% gain) HTTP faults : 0 net errors, 0 proto errors, 0 retried, 0 drops TCP handshakes : 351 total (152.0 req/conn) TCP faults : 0 failures, 0 timeouts, 3 purged External links : 202 skipped Reqs pending : 18692 Database statistics: Pivots : 880 total, 462 done (52.50%) In progress : 34 pending, 145 init, 221 attacks, 18 dict Missing nodes : 4 spotted Node types : 1 serv, 186 dir, 544 file, 16 pinfo, 76 unkn, 57 par, 0 val Issues found : 11 info, 75 warn, 57 low, 0 medium, 128 high impact Dict size : 263 words (263 new), 4 extensions, 256 candidates
At the end of the process, skipfish will dump all the data collected within the mediawiki_dir directory (defined by the -o option), that in turn contains an HTML file (index.html) which allows to view the report generated.
In the previous outcome, skipfish has only found out severe problems related to HTTP PUTs accepted.
So as to be able to understand the results offered by skipfish and if you do not have deep knowledge about web security (like me), you might take a look at the Browser Security Handbook, written and maintained by the same author who is developing skipfish.
Other interesting parameter is for example -A, used for passing HTTP authentication credentials.
And finally, also point out that you can tune skipfish in networking or crawling scopes, through different options which allow to set up for instance parameters related to TCP connections or the depth of the analysis. For getting more information you can check the project documentation.