Get the latest tech news

China’s DeepSeek Coder becomes first open-source coding model to beat GPT-4 Turbo


DeepSeek Coder V2 is being offered under a MIT license, which allows for both research and unrestricted commercial use.

Founded last year with a mission to “unravel the mystery of AGI with curiosity,” DeepSeek has been a notable Chinese player in the AI race, joining the likes of Qwen, 01.AI and Baidu. The original DeepSeek Coder, with up to 33 billion parameters, did decently on benchmarks with capabilities like project-level code completion and infilling, but only supported 86 programming languages and a context window of 16K. When tested on MBPP+, HumanEval, and Aider benchmarks, designed to evaluate code generation, editing and problem-solving capabilities of LLMs, DeepSeek Coder V2 scored 76.2, 90.2, and 73.7, respectively — sitting ahead of most closed and open-source models, including GPT-4 Turbo, Claude 3 Opus, Gemini 1.5 Pro, Codestral and Llama-3 70B.

Get the Android app

Or read this on Venture Beat

Read more on:

Photo of China

China

Photo of GPT-4 Turbo

GPT-4 Turbo

Photo of DeepSeek Coder

DeepSeek Coder

Related news:

News photo

China's Big Tech companies taught Asia to pay by scanning QR codes, but made a mess along the way

News photo

How innovative is China in nuclear power?

News photo

America's Defense Department Ran a Secret Disinfo Campaign Online Against China's Covid Vaccine