I spent the whole day yesterday testing DeepSeek, solving coding problems with Open Hands (previously called Open Devin).
The model feels rock solid. At times, it would go off track, but a simple reset of the window was enough to bring it back in line. Once that was done, it was smooth again.
DeepSeek has really raised the bar. What do you all think?
Yeah, DeepSeek really got me excited again about AI. The model is smart, and the cost makes it easy to add AI into projects without worrying about the budget. I had an idea for an AI-driven video game when ChatGPT first came out, and now it finally seems possible.
Kip said: @Rafe
You mean cheap APIs? Because with 685B parameters, not many people can run it locally.
Yeah, APIs. I haven’t compared prices much, but I tried DeepSeek through OpenRouter, and it was fast, smart, and super cheap. I used it for a while and only spent 5 cents on compute.
@Raven
Honestly, it depends on how you use it. If you like flexibility, OpenRouter might be the way to go. If you just want DeepSeek, going direct could be easier.
I find it weaker than Claude, but I don’t use it for coding. Surprised it’s getting so much hype.
I just chat with AI about different topics. I’ve tried 4o, Sonnet 3.5, all Gemini versions, Grok, and a bunch of open-source models. DeepSeek is better than most open-source ones, but I don’t think it’s on par with Sonnet or 4o.
It gets stuck in loops sometimes, ignores my prompts, and gives weird responses. Maybe it’s optimized for coding? I’ve used both the DeepSeek chat interface and OpenRouter.
Zev said: @Sterling
That could be it. I haven’t used it for regular conversation, only for coding.
I actually like using the real-time Gemini API for chat.
Same. I use the multimodal Gemini API more than ChatGPT’s voice mode. The only issue is the 15-minute limit. Gemini 2.0 follows instructions better than most models, especially for roleplay.
Noor said: @Oli
Even at full price, running it locally is still way more expensive than using the API.
It’s a mixture of a big model and a MoE (Mixture of Experts) model. It activates about 37B parameters per response. Running that on a CPU is possible but painfully slow.