How Website

Check out our free course!

We assume you already generally understand how the internet and websites work. If not, read these resources over.

Error Codes

Google Hacking

What is Google hacking (or Google dorking)? Well, Google (and other search engines) are built around a large number of web-crawlers that are constantly scraping the internet. They scoop up everything that is publicly available and index it, allowing you to search it. However... what happens when the web crawlers find something that they shouldn't?

Play around with a few of these Google hacks on Exploit-DB:

In order to prevent Google and other search engines from indexing a site, many sites use a robots.txt file to "block" the web crawlers. Robots.txt files are not actually part of an RFC yet, despite being an accepted part of the internet for years, but Google has recently proposed the Robots Exclusion Protocol as an official standard under Internet Engineering Task Force . Importantly, robots.txt are meant to be a good faith request by site administrators, and are regularly ignored by web crawlers. This means that any web crawler, or you, can visit a site's robots.txt and quickly identify any part of the site that the owner does not want crawled.

This often comes up in CTF problems, so it is always worth a quick check.

Visit the course page!

Hoppers Roppers 2020            Date: 2020-06-10 23:03:16

results matching ""

    No results matching ""