Back
Todo
Time-non-critical agents are super slow because you use flex mode in OpenAI to save costs. Now with gpt-5 they became even slower (and cheaper, so it's a good thing!). So you *need* to parallelize your agents. This is what I've been doing last night. My library #monoflux's flatMap method on a Flux (a Flux is an asyncronous array) can now handle () => Flux callbacks. It is now my number 1 engine to run *tons* of agents in parallel.
Home
Search
Messages
Notifications
More