atextcrawler crawls and indexes selected websites. It starts from a few seed sites and follows their external links. Criteria defined in plugin code determine which linked sites (and which of their resources) are (recursively) added to the pool.
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
ibu b7f3e174db Remove unwanted print 4 months ago
doc Make minimum text length configurable and actually remove elasticsearch documents 5 months ago
src/atextcrawler Remove unwanted print 4 months ago
tests Put under version control 6 months ago
.gitignore Put under version control 6 months ago
.pre-commit-config.yaml Put under version control 6 months ago
Pipfile Put under version control 6 months ago
Pipfile.lock Put under version control 6 months ago
README.md Put under version control 6 months ago
license.txt Put under version control 6 months ago
pyproject.toml Put under version control 6 months ago

README.md

atextcrawler is an asynchronous webcrawler indexing text for literal and semantic search.

Its client-side counterpart is atextsearch

atextcrawler crawls and indexes selected websites. It starts from a few seed sites and follows their external links. Criteria defined in plugin code determine which linked sites (and which of their resources) are (recursively) added to the pool.

atextcrawler is written in Python, runs a configurable number of async workers concurrently (in one process), uses tensorflow for embedding (paragraph-sized) text chunks in a (multi-)language model and stores metadata in PostgreSQL and texts in elasticsearch.