Skip to content(if available)orjump to list(if available)

DeepSeek-v3.1-Terminus

DeepSeek-v3.1-Terminus

13 comments

·September 22, 2025

storus

I tried V3.1 but it was driving me crazy by ignoring parts of user input, which R1 never did. I had many such instances when e.g. asking about running DeepSeek 671B it instead picked DeepSeek 67B because 671B is too large to exist so I must have made a mistake etc. I concluded that despite being better in benchmarks than R1, it was essentially useless due to this characteristics and I instead started using R1 at OpenRouter. Not sure why deepseek.com removed R1 and left only V3.1 without any ability to switch back, I guess it's cheaper to run.

esafak

Notable performance improvement in agentic tool use: https://huggingface.co/deepseek-ai/DeepSeek-V3.1-Terminus

yu3zhou4

I see no article in the link, just "news250922" header with some layout

sbinnee

> What’s improved? Language consistency: fewer CN/EN mix-ups & no more random chars.

It's good that they made this improvement. But is there any advantages at this point using DeepSeek over Qwen?

coder543

They seem fairly competitive with each other. You would have to benchmark them for your specific use case.

IgorPartola

I wish there was some easy resource to keep up with the latest models. The best I have come up with so far is asking one model to research the others. Realistically I want to know latest versions, best use case, performance (in terms of speed) relative to some baseline, and hardware requirements to run it.

exe34

> asking one model to research the others.

that's basically choosing are random with extra steps!

throwup238

Research not spit out the answer based on weights. Just ask Gemini/Claude to do deep research on /r/LocalLLama and HN posts.

comrade1234

MIT license that lets you run it on your own hardware and make money off of it.

coder543

Qwen3 models (including their 235B and 480B models) use the Apache-2.0 license, so it’s not like that’s a big difference here.