Posted by u/xandep
I regret ever finding LocalLLaMA
It all started with using "the AI" to help me study for a big exam. Can it make some flashcards or questions? Then Gemini. Big context, converting PDFs, using markdown, custom system instruction on Ai Studio, API. Then LM Studio. We can run this locally??? Then LocalLLama. Now I'm buying used MI50s from China, quantizing this and that, squeezing every drop in REAP, custom imatrices, llama forks. Then waiting for GLM flash, then Qwen, then Gemma 4, then "what will be the future of Qwen team?". Exam? What exam? In all seriousness, i NEVER thought, of all things to be addicted to (and be so distracted by), local LLMs would be it. They are very interesting though. I'm writing this because just yesterday, while I was preaching Qwen3.5 to a coworker, I got asked what the hell was I talking about and then what the hell did I expected to gain from all this "local AI" stuff I talk so much about. All I could thought about was that meme. https://preview.redd.it/o7e97f302aog1.png?width=932&format=png&auto=webp&s=98e0f8f9bd30bb9c49c18e3b7ed03751d605cc86
More from r/LocalLLaMA
Anthropic: "We’ve identified industrial-scale distillation attacks on our models by DeepSeek, Moonshot AI, and MiniMax." 🚨
I feel personally attacked
Distillation when you do it. Training when we do it.