|
Esri Geoportal Server 1.2.9
|
||||||||
PREV PACKAGE NEXT PACKAGE | FRAMES NO FRAMES |
See:
Description
Interface Summary | |
---|---|
Access | Access. |
Bots | Represents access policy from a single "robots.txt" file. |
MatchingStrategy | Matching strategy. |
WinningStrategy | Winning strategy. |
Class Summary | |
---|---|
BotsParser | Parser of "robots.txt" file. |
BotsUtils | Robots txt utility class/shortcut methods |
Enum Summary | |
---|---|
BotsMode | RobotsTxt mode. |
Directive | Robots.txt file directive. |
Provides support for reading and parsing robots.txt
Standard for Robots Exclusion is a mechanism allowing servers to communicate with web crawlers about it's access policy. This implementation follows recommendations found in the following sources:
http://www.robotstxt.org/orig.html
http://www.robotstxt.org/norobots-rfc.txt
https://en.wikipedia.org/wiki/Robots_exclusion_standard
https://developers.google.com/webmasters/control-crawl-index/docs/robots_txt
Behavior of the "robots.txt" mechanism can be configured through the following parameters in the gpt.xml configuration file:
bot.robotstxt.enabled: use of robots.txt during harvesting is enabled. Default: true
bot.robotstxt.override: allows user to override bot.robotstxt.enabled. Default: true;
bot.agent: name of the user agent for interpreting content of robots.txt. Default: "GeoportalServer"
Bots
,
BotsParser
,
BotsUtils
|
Esri Geoportal Server 1.2.9
|
||||||||
PREV PACKAGE NEXT PACKAGE | FRAMES NO FRAMES |