/AWS1/CL_BDAWEBCRAWLERCONF¶
The configuration of web URLs that you want to crawl. You should be authorized to crawl the URLs.
CONSTRUCTOR¶
IMPORTING¶
Optional arguments:¶
io_crawlerlimits TYPE REF TO /AWS1/CL_BDAWEBCRAWLERLIMITS /AWS1/CL_BDAWEBCRAWLERLIMITS¶
The configuration of crawl limits for the web URLs.
it_inclusionfilters TYPE /AWS1/CL_BDAFILTERLIST_W=>TT_FILTERLIST TT_FILTERLIST¶
A list of one or more inclusion regular expression patterns to include certain URLs. If you specify an inclusion and exclusion filter/pattern and both match a URL, the exclusion filter takes precedence and the web content of the URL isn’t crawled.
it_exclusionfilters TYPE /AWS1/CL_BDAFILTERLIST_W=>TT_FILTERLIST TT_FILTERLIST¶
A list of one or more exclusion regular expression patterns to exclude certain URLs. If you specify an inclusion and exclusion filter/pattern and both match a URL, the exclusion filter takes precedence and the web content of the URL isn’t crawled.
iv_scope TYPE /AWS1/BDAWEBSCOPETYPE /AWS1/BDAWEBSCOPETYPE¶
The scope of what is crawled for your URLs.
You can choose to crawl only web pages that belong to the same host or primary domain. For example, only web pages that contain the seed URL "https://docs.aws.amazon.com/bedrock/latest/userguide/" and no other domains. You can choose to include sub domains in addition to the host or primary domain. For example, web pages that contain "aws.amazon.com" can also include sub domain "docs.aws.amazon.com".
iv_useragent TYPE /AWS1/BDAUSERAGENT /AWS1/BDAUSERAGENT¶
Returns the user agent suffix for your web crawler.
iv_useragentheader TYPE /AWS1/BDAUSERAGENTHEADER /AWS1/BDAUSERAGENTHEADER¶
A string used for identifying the crawler or bot when it accesses a web server. The user agent header value consists of the
bedrockbot, UUID, and a user agent suffix for your crawler (if one is provided). By default, it is set tobedrockbot_UUID. You can optionally append a custom suffix tobedrockbot_UUIDto allowlist a specific user agent permitted to access your source URLs.
Queryable Attributes¶
crawlerLimits¶
The configuration of crawl limits for the web URLs.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_CRAWLERLIMITS() |
Getter for CRAWLERLIMITS |
inclusionFilters¶
A list of one or more inclusion regular expression patterns to include certain URLs. If you specify an inclusion and exclusion filter/pattern and both match a URL, the exclusion filter takes precedence and the web content of the URL isn’t crawled.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_INCLUSIONFILTERS() |
Getter for INCLUSIONFILTERS, with configurable default |
ASK_INCLUSIONFILTERS() |
Getter for INCLUSIONFILTERS w/ exceptions if field has no va |
HAS_INCLUSIONFILTERS() |
Determine if INCLUSIONFILTERS has a value |
exclusionFilters¶
A list of one or more exclusion regular expression patterns to exclude certain URLs. If you specify an inclusion and exclusion filter/pattern and both match a URL, the exclusion filter takes precedence and the web content of the URL isn’t crawled.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_EXCLUSIONFILTERS() |
Getter for EXCLUSIONFILTERS, with configurable default |
ASK_EXCLUSIONFILTERS() |
Getter for EXCLUSIONFILTERS w/ exceptions if field has no va |
HAS_EXCLUSIONFILTERS() |
Determine if EXCLUSIONFILTERS has a value |
scope¶
The scope of what is crawled for your URLs.
You can choose to crawl only web pages that belong to the same host or primary domain. For example, only web pages that contain the seed URL "https://docs.aws.amazon.com/bedrock/latest/userguide/" and no other domains. You can choose to include sub domains in addition to the host or primary domain. For example, web pages that contain "aws.amazon.com" can also include sub domain "docs.aws.amazon.com".
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_SCOPE() |
Getter for SCOPE, with configurable default |
ASK_SCOPE() |
Getter for SCOPE w/ exceptions if field has no value |
HAS_SCOPE() |
Determine if SCOPE has a value |
userAgent¶
Returns the user agent suffix for your web crawler.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_USERAGENT() |
Getter for USERAGENT, with configurable default |
ASK_USERAGENT() |
Getter for USERAGENT w/ exceptions if field has no value |
HAS_USERAGENT() |
Determine if USERAGENT has a value |
userAgentHeader¶
A string used for identifying the crawler or bot when it accesses a web server. The user agent header value consists of the
bedrockbot, UUID, and a user agent suffix for your crawler (if one is provided). By default, it is set tobedrockbot_UUID. You can optionally append a custom suffix tobedrockbot_UUIDto allowlist a specific user agent permitted to access your source URLs.
Accessible with the following methods¶
| Method | Description |
|---|---|
GET_USERAGENTHEADER() |
Getter for USERAGENTHEADER, with configurable default |
ASK_USERAGENTHEADER() |
Getter for USERAGENTHEADER w/ exceptions if field has no val |
HAS_USERAGENTHEADER() |
Determine if USERAGENTHEADER has a value |