TAG
— BROWSING TAG —

#DEEPSEEK

10
POSTS

DeepSeek V4: 1M‑Token Context and Budget Frontier AI Models

DeepSeek V4 launches with 1M‑token context, MoE‑based Pro and Flash models, and prices far below U.S. rivals, focusing on code and agent workloads.

DeepSeek V4 vs Claude/GPT: Strengths, Controversies Ahead

DeepSeek V4 eyes early March drop, topping Claude/GPT in code benchmarks with 1M context, but Blackwell chip scandal brews US ire.

Why US platforms are turning to Chinese AI models

Chinese open source AI like DeepSeek and Qwen is winning US clients by mixing solid performance with flexibility and lower operating costs.

How DeepSeek Trains Powerful Models On A Budget

The mHC idea lets DeepSeek train 3–27B models more efficiently, supporting its push to rival US labs while spending only a fraction on hardware.

Chinese Coding Models GLM 4.7 and M2.1 Step Up

Chinese coding models now sit close to Claude tier performance, while Doubao and Yuanbao prove domestic AI platforms can win huge mainstream audiences...

AI and Job Losses: DeepSeek Predicts Massive Workplace Shift

AI could take over most jobs in 10–20 years, warns DeepSeek’s researcher, urging tech firms to protect society amid rising risks and disruption.

DeepSeek R1: $294K training cost, ultra‑low inference rates

DeepSeek’s R1 cost ~$294K to train with 512 H800 GPUs and 80h. Inference costs are ~$0.55 input / $2.19 output per million tokens.

DeepSeek-V3 Unveiled: 236B Parameters and 128K Context for Free

DeepSeek-V3 majorly upgrades from V2 with a massive 236B parameter count, a vastly larger 128K context window, and superior performance, all offered f...

DeepSeek Open-Sources Nano-VLLM

DeepSeek's Nano-VLLM cuts VLLM's memory by 5x using dynamic KV cache, runs on CPU via NumPy, and adds speculative prefill. Targets edge AI devs.

DeepSeek has unveiled Distilled-R1

DeepSeek's R1 is a lightweight AI model that runs on a single GPU.

10 / 10

All 10 posts loaded