r/LocalLLaMA
Posted by u/__Maximum__
My gpu poor comrades, GLM 4.7 Flash is your local agent
Tools 442 points
150 comments
Yesterday
I tried many MoE models at 30B or under and all of them failed sooner or later in an agentic framework. If z.ai is not redirecting my requests to another model, then GLM 4.7 Flash is finally the reliable (soon local) agent that I desperately wanted. I am running it since more than half an hour on opencode and it produced hundreds of thousands tokens in one session (with context compacting obviously) without any tool calling errors. It clones github repos, it runs all kind of commands, edits files, commits changes, all perfect, not a single error yet. Can't wait for GGUFs to try this locally.
More from r/LocalLLaMA
r/LocalLLaMA · u/EmPips
New
Hot
My story of underestimating /r/LocalLLaMA's thirst for VRAM
Tools
1.3K 88 5 days ago
r/LocalLLaMA · u/Dark_Fire_12
New
zai-org/GLM-4.7-Flash · Hugging Face
Tools
719 225 Yesterday
r/LocalLLaMA · u/Fear_ltself
New
NVIDIA's new 8B model is Orchestrator-8B, a specialized 8-billion-parameter AI designed not to answer everything itself, but to intelligently manage and route complex tasks to different tools (like web search, code execution, other LLMs) for greater efficiency
I’ve seen some arguments we’ve reached AGI, it’s just about putting the separate pieces together in the right context....
Tools
705 129 6 days ago