mirror of
https://github.com/knightcrawler-stremio/knightcrawler.git
synced 2024-12-20 03:29:51 +00:00
Big rewrite - distributed consumers for ingestion / scraping(scalable) - single producer written in c#.
Changed from page scraping to rss xml scraping Includes RealDebridManager hashlist decoding (requires a github readonly PAT as requests must be authenticated) - This allows ingestion of 200k+ entries in a few hours. Simplifies a lot of torrentio to deal with new data
This commit is contained in:
@@ -6,4 +6,10 @@
|
||||
```
|
||||
docker-compose up -d
|
||||
```
|
||||
Then open your browser to `127.0.0.1:7000`
|
||||
Then open your browser to `127.0.0.1:7000`
|
||||
|
||||
If you'd like to enable crawling of RealDebridManager's shared hashlists which will massively boost your database cached entries,
|
||||
enter a readonly github personal access token in 'env/producer.env' as the 'GithubSettings__PAT=<token_here>' value.
|
||||
|
||||
You can scale the number of consumers, by changing the consumer deploy replica count in the compose file on line 87. This is currently set to 3.
|
||||
If you'd like to adjust the number of concurrent processed ingestions per consumer, thats the job concurrency setting within 'env/consumer.env'.
|
||||
Reference in New Issue
Block a user