Deepseek V3.1 Code: This REALLY THROWS Claude Code into the TRASH! (+V3.1 Test Results)

πŸš€ Add to Chrome – It’s Free - YouTube Summarizer

Category: Tech Review

Tags: AICodingDeepseekModelTech

Entities: ADER leaderboardsDeepseek V3.1Microsass FastRequestySonnetTMDB API

Building WordCloud ...

Summary

    Introduction
    • Deepseek V3.1 has been launched with major improvements.
    • The model is a hybrid reasoning model for both reasoning and non-reasoning tasks.
    Model Improvements
    • Better tool calling support, scoring higher in benchmarks.
    • Increased context limit from 65K to 128K.
    • Improved performance for both general and reasoning models.
    • Non-reasoning model excels in ADER leaderboards and is cost-effective.
    Usage and Integration
    • Deepseek can be a drop-in replacement for Sonnet.
    • Anthropic API format provided for easy integration.
    • Instructions on setting up environment variables and using Requesty for configuration.
    • Example of creating a movie tracker app using Deepseek.
    Conclusion
    • Deepseek V3.1 addresses previous issues and is recommended for coding tasks.
    • Non-reasoning variant is highly effective and cost-efficient.

    Transcript

    00:00

    [Music] [Applause] Hi, welcome to another video. So, Deepseek V3.1 was launched, but the major info about it wasn't available before.

    Now, it is available, and the

    00:17

    TLDDR is that it is a hybrid reasoning model that can basically do both reasoning and non-reasoning tasks. This allows for efficient performance.

    It also now has much better tool calling support which was one of the major pain

    00:34

    points of Deepseek R1 and V3. It now scores way better in multiple tool calling benchmarks and works much better in things like clin rue and open code.

    Similarly, they have also increased the

    00:49

    context limit to 128K from 65K which is a great upgrade as well. The model endpoints still exist the same way which is deepseek chat for the general model and the reasoning endpoint

    01:05

    for the reasoning variant instead of anything like reasoning effort. The performance is really improved here on my benchmark.

    The model is way better now even without reasoning. It can create floor plans pretty well though

    01:21

    they are not the best. The oneshot generation that just works for a model this cheap is awesome.

    The SVG Panda with the burger is one of the best I have seen yet. It is actually usable, which is not even something that Opus

    01:36

    can make. The autoplay chess also works pretty well.

    It doesn't always make legal moves, but it is still better than many other generations. The flying butterfly is also kind of cool.

    And overall this stuff is really

    01:52

    good. If we talk about the reasoning variant then it is also good but not as good as the non-reasoning variant at least for coding.

    So yeah it's not as great at least for coding but the non-reasoning model is

    02:08

    actually amazing in ADER leaderboards. It also comes close to opus for way cheaper pricing which is quite awesome for an open and cheaper model.

    It's really good and I think it is now one of the best options

    02:24

    to be a drop-in replacement for sonnet and that's what DeepS also thinks because they themselves now provide an anthropic API format along with details to make it a drop-in replacement in claw code and use this model. I've been using

    02:42

    it and it is really awesome. It feels a lot like Sonnet but it's 20x cheaper.

    So yeah, that's a great deal. Let me show you how you can also do the same and use it in interesting ways.

    But before we do

    02:57

    that, let me tell you about today's sponsor, Microsass Fast. Dreaming of launching a Micros or AI side project, but wasting weeks setting up O payments and SEO?

    Check out Microsass Fast, a next.js boilerplate with clerk, stripe, resend, hostgsql, and AI instructions

    03:15

    that cut hallucinations by 90% for vibe coding. Easy back-end integration with Python, Node, and Go.

    It is built and used by a CTO who helped 50 plus founders to launch SAS in the past year. You can save 50 plus hours and actually ship faster.

    Check now. Link is in the

    03:31

    description. Now, back to the video.

    First, make sure that you have Claude code installed and upgraded to the latest version. You can do either of them with the same command.

    Once that is done, you can export the

    03:47

    environment variables for base URL as Deepseek's anthropic base URL along with O token model and also the small fast model as deepseeek chat model. Also, open router doesn't support

    04:03

    anthropic format. So you can use claude code router for that or a better option that I use is requesty as it is way better than open router in terms of configuration and customization.

    It also comes with anthropic compatible

    04:21

    format and you should be able to use almost any model from there. I am going to be using deepsek from there but you can use the original API as well.

    For Requesty, you just have to change the base URL to support Requesty as

    04:38

    written in their docs and then just add deepsee slash in front of the DeepSeek chat endpoint and it should work. It actually works pretty well.

    The slash commands here all work fine along with MCPS which also work fine. For example,

    04:57

    I use the bite rover MCP here because I really like it for a memory layer. It allows you to plug it into all kinds of coders and it can create memories that are synced across all your coders.

    You can also share it with your teammates

    05:12

    which allows you to build rules and memories that stick to your projects. So yeah, this is something that I use and you can easily get it configured on cloud code.

    Just do that and you'll get some better results.

    05:28

    For example, here I have it started in a Nex.js project. I'm going to ask it to make me a simple movie tracker app that uses the TMDB API for the database.

    And what you'll see is that it actually goes ahead and works amazingly well. It can

    05:46

    do most of the tool calling without failing, which was a big downside in the previous version, but now it is amazingly better and actually works on par with Sonnet while being 20 times cheaper, literally. So yeah, it's

    06:02

    awesome. It can also use the to-do list stuff as well as MCPs without any issues.

    It's really great. for planning.

    You can change the model to Deepseek's reasoning variant as well if you wish to do that.

    06:18

    Anyway, the speed of the model is also now way better. It's probably because they don't need to deploy two models anymore and just one works, which produces way better speeds now and works well.

    So, it's pretty usable now.

    06:35

    Anyway, in a bit it gets it done. And you can see that it did this well without any major errors.

    And we can now go ahead and use it. It looks pretty great.

    It's on par with what Sonnet makes in one shot while being a lot

    06:51

    cheaper and pretty similar. So yeah, this makes a lot of sense to use.

    I prefer it over GLM for replacing and clawed as it's pretty great now. I was thinking Deepseek V3 3.1 was not as good at first, but as I used it, it felt like

    07:08

    Deepseek actually addressed almost all the major pain points of the original model, like tool calling, MCP support, and everything. It now works way better and is also way better at coding.

    So yeah, I'd highly recommend you check

    07:25

    this out. It's an awesome model.

    I don't know what happened with the reasoning variant. They probably just have that for things like mathematics because the non-reasoning variant now works amazingly well for coding and there's

    07:40

    probably no need to use sonnet for trivial tasks which saves a ton of money and costs. I really like it.

    Go ahead and check it out and use it all you want. Overall, it's pretty cool.

    Anyway, share your thoughts below and subscribe

    07:56

    to the channel. You can also donate via super thanks option or join the channel as well and get some perks.

    I'll see you in the next video. Bye.

    [Music]