Skip to main content
DeepSeekApple Siliconлокальный ИИ

DeepSeek 4 Flash q2 on M5: What a Local Test Run Revealed

Practical tests of DeepSeek 4 Flash q2 on an M5 MacBook with 128GB RAM show real-world performance: around 30 tok/s, up to 80GB memory usage, and some issues with tool calling breaking tags. For local AI implementation, this provides a clear benchmark on hardware needs and current limitations.

Technical Context

I love news like this not for the hype, but because it quickly grounds AI implementation in reality. It's simple: DeepSeek 4 Flash q2 is already being run locally on M5 MacBooks with 128GB of RAM, and live tests show around 30 tok/s.

For a single-user, local scenario, this is no longer a toy. Especially if you're looking into AI automation without the cloud, using private data with predictable latency.

What really caught my attention: DeepSeek itself uses up to 80GB of memory. The rest is consumed by adjacent processes like Claude Code, Codex, and other tools, which can easily take another 35GB.

So, this isn't just about the model but the entire work stack around it. On paper, you have 128GB, but in reality, that buffer disappears quickly if you don't keep the machine almost dedicated to inference.

Another real-world nuance: tool calling isn't perfect, and the model sometimes forgets to close tags. I consider these not cosmetic flaws but engineering details because they are what break agentic pipelines and automated action chains.

The good news is that this looks like a fixable problem at the wrapper, validation, and post-processing level. The bad news is you can't blindly rely on it out-of-the-box if your production logic depends on a strict format.

What This Means for Business and Automation

I see three practical takeaways here. First: deploying large models locally on Apple Silicon is now a realistic discussion, not just an experiment, for teams that value privacy and control.

Second: the hardware threshold hasn't gone away. If you don't have 128GB and discipline with background processes, the beautiful idea quickly turns into a battle for memory and an unstable UX.

Third: the winners are those who need a local code assistant, an internal agent, or private document processing. The losers are those expecting cloud-level speed and perfect tool use without additional engineering.

At Nahornyi AI Lab, we analyze these cases hands-on: where a local model is truly more cost-effective than an API, how to build an AI architecture without unnecessary costs, and how to safeguard tool calling so automation doesn’t fall apart over minor details. If you're considering a local AI automation pipeline, we can calmly assess your stack and build a solution without guesswork from forums.

Beyond optimizing specific models like DeepSeek for local hardware, understanding different local assistant implementations is crucial for practical applications. We previously explored Rust LocalGPT, which offers a single-binary local assistant with persistent memory and an HTTP API, showcasing another approach to practical AI implementation without the overhead.

Share this article