Information Gathering from Robots.txt File A website owner creates a robots.txt file to list the files or directories a web crawler should index for providing search results. Poorly written robots.txt files can cause the complete indexing of website files and directories. If confidential files and directories are indexed, an attacker may easily obtain information such as passwords, email addresses, hidden links, and membership areas. If the owner of the target website writes the robots.txt file without allowing the indexing of restricted pages for providing search results, an attacker can still view the robots.txt file of the site to discover restricted files and then view them to gather information. An attacker types URL/robots.txt in the address bar of a browser to view the target website’s robots.txt file. An attacker can also download the robots.txt file of a target website using the Wget tool.
Certified Ethical Hacker(CEH) Version 11 pg 1650
I would go with robots.txt. The question asks a file and with the content of the robots.txt the hacker can found directories which should be not visible.
Web Server Attack Methodolog
Information Gathering from Robots.txt File
The robots.txt file contains the list of the web server directories and files that the web site owner wants to hide from web crawlers. Poorly written robots.txt files can cause the complete indexing of website files and directories. If confidential files and directories are indexed, an attacker may easily obtain information such as passwords, email addresses, hidden links, and membership areas. (P.1650/1634)
The answer is B. Robots.txt.
It is called comprehensive reading people.
The question says which FILE.
Robots.txt is a file.
Documents Root is a Directory Folder NOT a file.
What is a robots txt file used for?
A robots. txt file tells search engine crawlers which URLs the crawler can access on your site. This is used mainly to avoid overloading your site with request. It can be used to discover the structure of a website during web-server footprinting.
According CEHv11 P.1650
An attacker can simply request the Robots.txt file from the URL and retrieve sensitive information such as
the root directory structure and content management system information about the target website
Correct answer is Document root:
The document root is a directory (a folder) that is stored on your host’s servers and that is designated for holding web pages.
Correct answer is Document root:
Explanation:
The document root is a directory (a folder) that is stored on your host’s servers and that is designated for holding web pages. When someone else looks at your web site, this is the location they will be accessing.
In order for a website to be accessible to visitors, it must be published to the correct directory, the “document root.”
You might think that there would only be one directory in your space on your host’s servers, but often hosts provide services beyond just publishing a website. In this case, they are likely to set up every account with several directories, since each service would require its own.
A voting comment increases the vote count for the chosen answer by one.
Upvoting a comment with a selected answer will also increase the vote count towards that answer by one.
So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.
blacksheep6r
Highly Voted 1Â year, 6Â months agotille
Highly Voted 1Â year, 11Â months agoDaniel8660
Most Recent 6Â months, 3Â weeks agodisil98445
11Â months agovolatile
11Â months agoEngnSu
11Â months agoMadhusudanan
1Â year agoalopezme
1Â year, 4Â months agoBot001
1Â year, 6Â months agobrdweek
1Â year, 8Â months agoANDRESCB1988
1Â year, 9Â months agobeowolf
1Â year, 10Â months agocerzocuspi
2Â years agoQuidProQuoo
1Â year, 11Â months agogenerate159357
1Â year, 8Â months agoamericaman80
2Â years ago_Storm_
2Â years ago