wget - Robot names for robots.txt -


suppose have website uses wget crawl other websites. provide website owners chance of not being crawled website. should use robot name wget in robots.txt file, or have create other name?

common practice disallow , allow popular uas this:

user-agent: google disallow:   user-agent: * disallow: / 

so think don't have problems using wget way


Comments

Popular posts from this blog

c++ - Function signature as a function template parameter -

algorithm - What are some ways to combine a number of (potentially incompatible) sorted sub-sets of a total set into a (partial) ordering of the total set? -

How to call a javascript function after the page loads with a chrome extension? -