llama-server auto-migrates cache to HuggingFace paths without user input. Following the ggml acquisition, infrastructure consolidation is accelerating—but silent breaking changes are testing developer patience. #LocalLLaMA #ggml
bymachine.news/llama-server-huggingface...
Qwen3.5 122B beats smaller Coder Next model on speed and accuracy. Better training, smarter optimization, superior quantization—not smaller = faster. Real-world performance > benchmark specs. #LocalLLaMA #Qwen
bymachine.news/qwen-switch-larger-model...
Local first Fill-in-the-Middle (FIM) with llama.cpp
> leaf.eagleusb.com/3mhv6sz2pf22b
#llm #localllama
3/
#LocalLLaMA, roleplay users and conservative testers love it because it doesn’t suffer from the “woke mindset” that plagues Claude (the worst), ChatGPT or Gemini.
Reason: trained on Chinese internet data + strict government control.
📰 Qwen3.5 Models Gain Traction for Performance, Efficiency
The Qwen3.5 series, particularly the 35B-A3B model, is gaining popularity in the LocalLLaMA community for its impressive pe...
www.clawnews.ai/qwen3-5-models-gain-trac...
#LocalLLaMA #Qwen35 #AIModels
#ai #llm #localllama #localllm #grownostr #nostr #gfy
Not your own locally hosted LLM? You are giving away your thoughts and ideas to corporations/governments, who pay for what and how you think.
Playing with compar:IA, the new LLM chatbot comparison arena created by the French government: https://comparia.beta.gouv.fr
I particularly like the "frugal" mode that allows you to compare two randomly chosen small/cheap models against each other! It's great for testing many different models […]
Audiobook Generator GUI that can clone your voice like 11Labs but hosted locally. Fun little project so I could listen to the books I've written and make sure everything sounded right.
github.com/Jeremy-Harpe...
#chatterbox
#audiobook
#author
#localllama
#audible
#audioAI
Haha, my Raspberry Pi 5 is actually faster at CPU-only inference than my old laptop. LocalScore 23, 9.5 tokens/seconf generation.
Indeed, the CPU-only performance is even worse. The LocalScore on the tiny 1B model is only 16, with a text generation speed of 7.7 tokens/second.
https://www.localscore.ai/result/235
Let's see if I can run this on a Raspberry Pi for comparison...
#LocalScore #llm #benchmark #LocalLlama
My hobby: running LocalScore.ai to benchmark how fast (ehm) my 2018 laptop runs a tiny 1B LLM. The laptop has a NVIDIA MX150 mobile GPU, 2GB VRAM. I guess it was intended for Photoshop filters or CAD stuff.
I got a LocalScore of 101 on the tiny model using the GPU (13.5 tokens/second for […]
Ok, so I'm running the new Orpheus TTS by canopy labs on llama.cpp and changing the top_p sampler to min_p gives me a 2x to 3x t/s speed boost. Why? Haven't seen this happen.
Any ideas?
#ai #localllama
Meta annonce 1 milliard de téléchargements pour Llama! 🦙✨ Mais est-ce vraiment "open source" ou juste "open weights"? La communauté #LocalLLaMA débat: délais trop longs entre versions, besoin de modèles plus accessibles. L'IA locale défie le cloud! #IALibre #TechQuébec patb.ca/r/e91
[1] Deepseek: Hold my beer. I'm going in!
#LocalLlama: Deepseek r1:1.5b running now
Talks in US congress about banning it ...so I figured it is worth a look. Besides, it is touted as the best at coding.
I downloaded on my average-spec'd laptop.
I followed these instructions:
Just a shout out to #ollama, great stuff!
#llama #localllama
Why DeepSeek’s web version is raising security alarms - Fast Company
www.fastcompany.com/91273103/chi...
#LocalLlama
Google has some ‘good ideas’ for putting ads in Gemini | Digital Trends
www.digitaltrends.com/computing/no...
#LocalLlama #LLM
To run your own Local AI Chatbot #LocalLLama , here is the minimum required hardware to work comfortably
Apple Mac Studio M1 Max 10 cœurs 64 Go RAM 1 To SSD (used price around $2000)
$2700 will get you the right specs with a new Mac Studio M2.
I’m waiting for upcoming M4 Mac Studio
Stay Free
#LocalLLama
What stops you from using #LocalLLama instead of paid and proprietary solutions like #ChatGPT?
Stay free with Local AI.
Good stuff ollama.com/library/qwq
What if Alibaba provided the most underrated AI model?
#LocalLLaMa
Build your own LLM-powered git commit message generator!🚀
Uses your local LLM using Ollama so your private data doesn’t leave your machine. 🔐
#GenerativeAI #LLM #ArtificialIntelligence #LocalLlama
youtu.be/YPeNoeVCWxo