Skip to content

Comments

Speculation on how sub-googlebots work, based on public documentation#2

Open
willcritchlow wants to merge 7 commits intomasterfrom
subgooglebot
Open

Speculation on how sub-googlebots work, based on public documentation#2
willcritchlow wants to merge 7 commits intomasterfrom
subgooglebot

Conversation

@willcritchlow
Copy link
Owner

If we pass in a comma-separated list of two user agents (more specific first, e.g. googlebot-image,googlebot) then run newly-exposed method that returns a response based on whether the first user agent is allowed / disallowed by a specific rule, and if not, falls back to a regular check on the second user agent.

Will Critchlow and others added 7 commits November 12, 2019 09:13
First should be the specific user agent, and second should be the general user agent whose ruleset the crawler should obey iif there are no specific rules targeting the first user agent
Expose tuple functionality in the interface to robots and call from the wrapper if two user agents passed in the string comma-separated.

Add test cases.
and documentation
One of the ways listed in the open source project for compiling and testing the project is using bazel - which creates directories that should be ignored by git as we never want to check them in
Ignore bazel build directories
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant