A small web crawler named aranea (Latin for spider). https://www.bananas-playground.net/projekt/aranea/
Banana a8bdf7dc15 do not use finish/( | hai 1 semana | |
---|---|---|
documentation | hai 1 semana | |
lib | hai 1 mes | |
storage | %!s(int64=2) %!d(string=hai) anos | |
.gitignore | hai 1 mes | |
CHANGELOG | hai 1 mes | |
COPYING | hai 1 mes | |
LICENSE | hai 1 mes | |
README.md | hai 1 semana | |
TODO | hai 1 semana | |
VERSION | %!s(int64=2) %!d(string=hai) anos | |
cleanup.pl | hai 1 semana | |
config.txt | hai 1 mes | |
fetch.pl | hai 1 semana | |
parse-results.pl | hai 1 semana | |
setup.sql | %!s(int64=2) %!d(string=hai) anos |
https://www.bananas-playground.net/projekt/aranea
A small web crawler named aranea (Latin for spider). The aim is to gather unique domains to show what is out there.
It starts with a given set of URL(s) and parses them for more URLs. Stores them and fetches them too. -> fetch.pl
Each URL result (Stored result from the call) will be parsed for other URLs to follow. -> parse-results.pl
After a run cleanup will gather all the unique Domains into a table. Removes URLs from the fetch table which are already enough. -> cleanup.pl
The table url_to_ignore
does have a small amount of domains and part of domains which will be ignored.
Adding a global SPAM list would be overkill.
A good idea is to run it with a DNS filter, which has a good blocklist.