Today I noticed the page of Baidu's search results, there is a row of relief, point open, and a bunch of "XXX does not assume any legal responsibility". Article 6 is "If any website does not want to be included by Baidu Online Network Technology (Beijing) Co., Ltd., it should be reacted to the service website or Baidu company, or in the website's page according to the rejection Spider Protocol Filling the tag of the reject, otherwise, Baidu's search engine will appreciate it as a collection of websites. "From here, you can learn about the search engine related law, the website is not clearly expressed, the refusal is the default is acceptable, no wonder Internet is open.
The feeling of European and American movies is that open women do not have a clear refusal. The real woman's experience is not a clear acceptance.
Prohibit search engine recordings
One. What is a robots.txt file?
The search engine automatically accesses web pages on the Internet through a program robot (also known as Spider).
You can create a plain text file robots.txt in your website, declare that the site does not want to be accessed by the Robot in this file, so that the part or all of the content can be included in the search engine, or Specifies that the search engine only includes the specified content.
II. Where is the robots.txt file?
Robots.txt files should be placed under the root directory of the website. For example, when Robots accesses a website (such as http://www.abc.com), first check if there is http://www.abc.com/robots.txt this file, if the robot finds This file will determine the scope of its access according to the content of this file.
Website URL The corresponding Robots.txt's URL http://www.w3.org/ http://www.w3.org/robots.txt http://www.w3.org:80/ http: // www. W3.org:80/robots.txt The following is Robots.txt in some famous sites:
http://www.cn.com/robots.txt
http://www.google.com/robots.txt
http://www.ibm.com/robots.txt
http://www.sun.com/robots.txt
http://www.eachnet.com/robots.txt
Three. Robots.txt file format
The "robots.txt" file contains one or more records that are separated by the space line (with Cr, Cr / NL, or NL as the end value), each record format is as follows:
"
In this document, you can use # to annotate, specific usage methods, and practices in UNIX. Records in this file typically start with a row or multi-line User-Agent, and there are several Disallow lines, and the details are as follows:
User-agent: The value of this item is used to describe the name of the search engine Robot. In the "Robots.txt" file, if there are multiple User-Agent records, multiple Robot will be subject to the limit, for the file Say, there must be at least one User-Agent record. If the value of this item is *, the protocol is valid for any machine, in the "Robots.txt" file, "user-agent: *" records can only have one. Disallow: The value of this item is used to describe a URL that does not want to be accessed, which can be a complete path, or some, any URL starting with Disallow is not accessed by Robot. For example, "disallow: / help" does not allow search engine access to /Help.html and /Help/index.html, and "Disallow: / Help /" allows Robot to access /Help.html, not access / help / index .html. Any disallow record is empty, indicating that all parts of the site allow access, in the "/ ROBOTS.TXT" file, at least one DisliW record. If "/Robots.txt" is an empty file, the site is open for all search engines Robot.
Four. Robots.txt file usage example
Here are some of the basic usage of Robots.txt:
l Disable all search engines from accessing website: user-agent: * dispialow: /
l Allow all Robot to access user-agent: * disallow: or you can also build an empty file "/Robots.txt" file
l Support all the sections of all search engines to access the website (CGI-BIN, TMP, PRIVATE directory in the following example) User-agent: * dispialow: / cgi-bin / disallow: / tmp / disallow: / privat /
l Disable a search engine access (BADBOT in the following example) user-agent: BadbotdisAllow: /
l Only a search engine is allowed (Webcrawler) user-agent: WebcrawlerDisAllow: user-agent: * dispialow: user-agent: * disallow: user-agent: * dispialow: /
The following gadgets specifically check the validity of the robots.txt file:
http://www.searchengineworld.com/cgi-bin/robotcheck.cgi
V. Robots Meta label
1, what is a Robots Meta label
The robots.txt file is mainly to limit the search engine access to the entire site or directory, while the Robots Meta label is primarily for one specific page. As with other Meta tags (such as using language, page description, keywords, etc.), Robots Meta tags are also placed in the
head> of the page, specifically to tell the search engine Robot how to capture the page Content. Specific form is similar (see the black body portion):
hEAD>
...
body>
html>
2, Robots Meta labeling:
There is no case in the Robots Meta tag, name = "robots" means all search engines, which can be written to Name = "baiduspider" for a specific search engine. The Content section has four command options: Index, NoIndex, Follow, NOFOLLOW, and Dances are separated.
The Index instruction tells the search robot to grab the page;
The FOLLOW instruction indicates that the search robot can continue to capture along the link on the page;
The default value for the Robots Meta tag is index and follow, except for INKTOMI, for it, default is index, nofollow.
In this way, there are four combinations:
among them
can be written
;
It should be noted that the above-mentioned Robots.txt and Robots Meta tags restricting the search engine robot (robots) to grab the site content is just a rule, and you need to search for engine robots. It is not every Robots comply.
At present, the vast majority of search engine robots comply with Robots.txt rules, and for the Robots Meta label, there is not much support, but it is gradually increased, such as the famous search engine Google fully supports, and Google also increases. A directive "Archive" can limit whether Google retain web snapshots. E.g:
Represents to capture the page in this site and link to the page, but the page snapshot of the page is not retained on GoOLGE.