The web, much just like the globe, includes a dark side to that. And by dark side, it’s not simply adult websites that we’re talking regarding, however one thing even larger and a lot of complicated. Just imagine that you are in the ship, your sonar detector can detect up to certain depth after that it can’t detect. The part which your ship can detect is deep web and the part which can’t be detected or takes time to detect called the dark web.
Deep web is that part of the web where people can’t access through their popular search engine. It is close to 500 times depth the size of the Web that we are familiar with. The deep web data — user databases, registration-required web forums, webmail pages, paywall-protected pages — are not indexed by search engines such as Google and Yahoo, and therefore cannot be found by simple search functions.
Dark web is much smaller than the deep web and is made up of all different kinds of websites that sell drugs, weapons and counterfeit currency. It is a virtual underworld where everything is anonymous and unmonitored, far, far-off from the prying eyes of intelligence and enforcement agencies across the world. Drug dealing, arms commercialism, hacking, harlotry, phishing, scams, frauds, the act of terrorism and plenty of such malicious activities are transacted anonymously via the dark web. All these transactions take place in dark web markets.
These hidden networks which avoid their presence on the Surface Web, and its URLs are tailed up with .onion. These .onion domains are not indexed by regular search engines, so people can only access Dark Web with special software called ‘The Onion Browser,’ referred to as TOR.
Why search engines can’t index these webs?
- Limited access content: Sites that limit access to their pages in an exceedingly technical approach.
- Non-HTML/text content: Matter content encoded in transmission (image or video) files or specific file formats not handled by search engines.
- Private Web: Sites that need registration and login.
- Scripted content: Pages that square measure solely accessible through links created by JavaScript moreover as content dynamically downloaded from net servers via Flash or mythical being solutions.
- Software: Bound content is by choice hidden from the regular net, accessible solely with special computer code, similar to Tor, I2P, or alternative darknet computer code. As an instance, Tor permits users to access websites victimization the .onion server address anonymously, activity their informatics address.
- Unlinked content: Pages that don’t seem to be coupled to by alternative pages, which can stop net locomotion programs from accessing the content. This content is stated as pages while not backlinks. Also, search engines don’t forever notice all backlinks from searched web content.
- Internet archives: Net repository services similar to the Wayback Machine alter users to examine archived versions of web content across time, together with websites that became inaccessible, and don’t seem to be indexed by search engines similar to Google.
- Contextual Web: Pages with content variable for various access contexts.
- Dynamic content: Dynamic pages that square measure came back in response to a submitted question or accessed solely through a type, particularly if open-domain input parts square measure used; such fields square measure onerous to navigate while not domain information.