by jhancock on 10/16/25, 6:29 AM
I've been using GLM-4.6 since its release this month. It's my new fav. Using it via Claude Code and the more simple Octofriend
https://github.com/synthetic-lab/octofriendHosting through z.ai and synthetic.new. Both good experiences.
z.ai even answers their support emails!! 5-stars ;)
by zozbot234 on 10/16/25, 7:33 AM
For those interested in building Ollama locally, note that as of a few hours ago, experimental Vulkan Compute support (will not be in official binary releases as of yet) has been merged on the github main branch and you can test it on your hardware!
by qqxufo on 10/16/25, 8:17 AM
Interesting to see more people mentioning GLM-4.6 lately — I’ve tried it briefly and it’s surprisingly strong for reasoning tasks. Curious how it compares to Claude 3.5 in coding throughput though?
by danans on 10/16/25, 4:44 PM
Question for those using local models for coding assistance: how well do the best locally runnable models (running on a laptop with a GPU) work for the easy case:
Writing short runs of code and tests after I give an clear description of the expected behavior (because I have done the homework). I want to save the keystrokes and the mental energy spent on bookkeeping code, not have it think about the big problem for me.
Think short algorithms/transformations/script, and "smart" auto complete.
No writing entire systems/features or creating heavily interpolated things due to underspecified prompts - I'm not interested in those.
by mike_d on 10/16/25, 7:06 AM
> For users with more than 300GB of VRAM, qwen3-coder:480b is also available locally.
I haven't really stayed up on all the AI specific GPUs, but are there really cards with 300GB of VRAM?
by esafak on 10/16/25, 2:17 PM
Has anybody that has tried their cloud product care to comment? How does it compare with Anthropic's and OpenAI's offerings in terms of speed and limits?
by bigyabai on 10/16/25, 7:09 AM
Been disappointed to see Ollama list models that are supported by the cloud product but not the Ollama app. It's becoming increasingly hard to deny that they're only interested in model inference just to turn a quick buck.
by skeeter2020 on 10/16/25, 2:27 PM
Seems appropriate that the top-level image is a ~sheep~ llama wearing a headband that says "coder"...
by qwe----3 on 10/16/25, 6:18 AM
Just a paste of llama.cpp without attribution.