r/LocalLLaMA
Posted by u/ayylmaonade
GLM 4.7 Flash official support merged in llama.cpp
Tools 352 points
61 comments
Yesterday
External link:
https://github.com/ggml-org/llama.cpp/pull/18936More from r/LocalLLaMA
r/LocalLLaMA · u/EmPips
New
Hot
My story of underestimating /r/LocalLLaMA's thirst for VRAM
Tools
1.3K 88 5 days ago
r/LocalLLaMA · u/Dark_Fire_12
New
zai-org/GLM-4.7-Flash · Hugging Face
Tools
719 225 Yesterday
r/LocalLLaMA · u/Fear_ltself
New
NVIDIA's new 8B model is Orchestrator-8B, a specialized 8-billion-parameter AI designed not to answer everything itself, but to intelligently manage and route complex tasks to different tools (like web search, code execution, other LLMs) for greater efficiency
I’ve seen some arguments we’ve reached AGI, it’s just about putting the separate pieces together in the right context....
Tools
705 129 6 days ago