Activity
Mon
Wed
Fri
Sun
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
Jan
Feb
Mar
What is this?
Less
More

Memberships

Clief Notes

16.1k members • Free

TC
This Community Moved

1.8k members • Free

AI Launchpad

21.6k members • Free

Chase AI Community

53.6k members • Free

Agent Zero

2.5k members • Free

Disrupter Academy

3.2k members • Free

The Build Room+

2.4k members • $97/month

1 contribution to Agent Zero
Huge token spend / API call volume
Hi guys. New to Agent Zero but familiar with what I’d consider ‘normal’ token usage based on experience with n8n, bolt, etc. After spending the day with A0 (using requesty for LLM APIs), I’m constantly seeing single messages eating over 100k tokens across 10+ api calls - which gets huge very quickly. I’ve tried a lot of potential solutions - from adding the ‘true’ flag for caching in the LiteLLM input field - to trying to control api call volume and token usage via global behaviour instructions and going down rabbit holes trying to sort something out via Perplexity at the VM level. Is this standard with A0? That is - it’s just a token hungry monster that sends full system prompts, verbatim chat thread copy, summarised chat thread copy and more in every API call associated with every single message? I’m just trying to wrap my head around this because it just doesn’t make sense. And that could simply be down to the fact that I’m a ln A0 noob. Any guidance appreciated 👍
Huge token spend / API call volume
1 like • Feb 22
@Justin Brown nice, thanks mate
1 like • Feb 24
@Mr Sonic thanks for the heads up. Would you mind sharing the provider / model setup you’ve settled on?
1-1 of 1
Leigh Rogers
1
1point to level up
@leigh-rogers-9191
Digital plumber and perpetual AI rabbit-hole faller.

Active 1d ago
Joined Feb 10, 2026
Powered by