I'm currently working on this.
I'll try to implement an auto-retry for whenever the rate-limit is reached but for now I've simply added two settings so you can customize the "request interval limit" for "api.twitter.com" (e.g. profile crawling) and "twimg.com" (downloading)".
Setting the api.twitter.com request interval limit to 2000-3000ms might already help to completely avoid ever reaching the API rate-limit.
I've also noticed that twitter seems to be very generous regarding such limits:
I was easily able to do 3000 requests with only 500ms of wait time in between without running into a limit but this may vary based on a lot of favtors such as logged in state (account/no twitter account), IP, country, ...
Once the limit is reached, twitter will return http-headers contaning a timestamp that defines when the limit will be over and the complete required wait time in seconds.
In my tests this was always 180 seconds.
I will try to modify future versions of our crawler in a way that auto-resume is possible.
A users' JD crashes and the first thing to ask is:
Quote:
Originally Posted by Jiaz
Do you have Nero installed?
That's true James
Quote:
Originally Posted by James
Die Leute verstehen einfach nicht dass nur weil man mit einer Waffe auch auf Menschen schießen kann dass ein Schützenverein kein Ort für Amoklaufideen ist
Last edited by pspzockerscene; 17.03.2022 at 14:45.
Reason: Improved readability