Skip to main content
локальные LLMtool callingAI automation

Small LLMs and Local Agents: Are We There Yet?

A recent benchmark tested small 3B-9B open models on coding, web scraping to JSON, and tool calling. This is crucial for businesses, clarifying where local AI automation is feasible within 4GB of VRAM and where cutting costs is risky. It shows the practical limits of on-device AI agents today.

Technical Context

I appreciate tests like these not for the fancy charts, but for the down-to-earth question they answer: can you build proper AI automation locally without buying a dedicated server for every minor task? This benchmark specifically tested small 3B-9B open-source models on three tasks you could confidently assign to a real developer.

The scenarios were practical: adding small features to a frontend and backend, finding data online, filtering it, saving it to JSON, and then separately testing tool calling. It's this third point where the whole "local agents on a budget" conversation usually falls apart.

The VRAM situation is encouraging. The discussion revealed that some of these models fit within a 4GB maximum, especially with 4-bit quantization. For 3B models, this is already a workable range, provided you don't inflate the context or layer on a heavy agentic loop with numerous tools.

Model-wise, I'd look at families like SmolLM3-3B, Gemma 3 4B, and some 7B-9B variants only if you're meticulously managing memory. For simple code and data processing, small models no longer feel like toys. However, their tool calling is still finicky. They handle simple tools well but quickly start hallucinating pathways in multi-step logic.

This is where I'd distinguish between "can call a function" and "can operate reliably within an agentic workflow." These are two very different benchmarks.

Impact on Business and Automation

The first takeaway is simple: local AI integration has become more realistic for narrow tasks. If you need to parse data, filter it, format it into JSON, perform minor developer operations, or create internal utilities, a small model under 4GB of VRAM can already be cheaper and more convenient than the cloud.

The second point is less pleasant: if your process relies on reliable tool calling, especially with multiple steps and result verification, deploying small models without a safety net is risky. I would add strict validators, retry logic, and routing to a more powerful model as a fallback.

Teams that need on-device operation, privacy, and low running costs are the winners here. Those hoping to replace a production-grade agent with a single "lightweight" model without proper engineering support will lose out.

At Nahornyi AI Lab, we solve these borderline problems for our clients: determining where a local model is sufficient and where a proper AI architecture with hybrid routing is necessary. If your processes are bogged down by manual routines or expensive API calls, my team and I can help you build an AI solution development plan without the magic wand and with clear economics.

As we explore the capabilities of small models in agentic workflows and tool use, it's crucial to also consider their inherent security challenges. We've previously covered how Unicode homoglyphs can deceive AI agents into phishing or executing malicious commands, a vital security guide for robust AI automation and tool use implementation.

Share this article