You are viewing a single comment's thread from:

RE: Curated by Thoth - 2025-08-18 20:53Z

in #thoth-test3 days ago

It looks like I've reached the end of the line for gemini-2.5-pro (free tier). I'm guessing that they deprioritize free API accounts when usage is high, and the failure rate has been very high this week with that model. It's not hitting rate limits, but the errors I'm seeing have the same impact.

Fortunately, it seems like gemini-2.5-flash does a decent job with the blog post format (so far, anyway), so I'll probably stick with that one unless/until I have some reason to switch.

Sort:  

What happens to Thoth when the free usage limit is exhausted? Does the script then stop sending summaries?

We had discussed whether AI could analyse all posts and suspected that it would not be feasible with free subscriptions. Now we have practical proof.

What happens to Thoth when the free usage limit is exhausted? Does the script then stop sending summaries?

I just have it scheduled to run twice per day, starting from a random block (or from a specified block, depending upon the config), so if the rate limit were exceeded, the runs would fail and abort until the next day. Or I would cancel the 2nd run if the first one failed. So far, it hasn't gotten close to the limits, though.

If multiple people were running Thoth, each would need to provide their own API keys, so they'd each have their own rate limit. I actually set up a paid API key for myself, too, just in case, but I haven't had to use it yet (and I don't plan to for the foreseeable future).

We had discussed whether AI could analyse all posts and suspected that it would not be feasible with free subscriptions. Now we have practical proof.

Right. One person definitely couldn't do it. I guess it's theoretically possible with decentralization. It would take a lot of people with their own individual free subscriptions, though. The most you can get from gemini is with the flash-lite model at 1000 requests per day or gemma at 14400 requests per day.

The ArliAI API is free with unlimited usage, but it only accepts one connection at a time, and it's really slow, so I don't think it would be able to analyze every post in the time available. Also, I haven't reworked the prompts for their gemma model after they pulled the rug out on the previous models that were in their free tier.