8000 GitHub - merwin-asm/Robots.TXT: Robots.txt parser for python || Better than the OG one for some reasons
[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to content

merwin-asm/Robots.TXT

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 

Repository files navigation

Robots.txt Parser

Robots.txt parser for python. Better than the OG one for some reasons.

What advantages you will have using this instead of OG 'urllib.robotparser' ?

  • Supports the use of full url / any url rel to the site
  • Supports proxies
  • Does show dissallowed sites

Use :

import robots_txt

res = disallowed("https://www.google.com/blablabla", proxy=None) 
# returns the disallowed sites for google.com

About

Robots.txt parser for python || Better than the OG one for some reasons

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

0