Tag: Deepseek v3
-
Hacker News: Has DeepSeek improved the Transformer architecture
Source URL: https://epoch.ai/gradient-updates/how-has-deepseek-improved-the-transformer-architecture Source: Hacker News Title: Has DeepSeek improved the Transformer architecture Feedly Summary: Comments AI Summary and Description: Yes **Summary:** The text discusses the innovative architectural advancements in DeepSeek v3, a new AI model that boasts state-of-the-art performance with significantly reduced training times and computational demands compared to its predecessor, Llama 3. Key…
-
Hacker News: Qwen2.5-Max: Exploring the Intelligence of Large-Scale Moe Model
Source URL: https://qwenlm.github.io/blog/qwen2.5-max/ Source: Hacker News Title: Qwen2.5-Max: Exploring the Intelligence of Large-Scale Moe Model Feedly Summary: Comments AI Summary and Description: Yes Summary: The text discusses the development and performance evaluation of Qwen2.5-Max, a large-scale Mixture-of-Expert (MoE) model pretrained on over 20 trillion tokens. It highlights significant advancements in model intelligence achieved through scaling…
-
Simon Willison’s Weblog: Anomalous Tokens in DeepSeek-V3 and r1
Source URL: https://simonwillison.net/2025/Jan/26/anomalous-tokens-in-deepseek-v3-and-r1/#atom-everything Source: Simon Willison’s Weblog Title: Anomalous Tokens in DeepSeek-V3 and r1 Feedly Summary: Anomalous Tokens in DeepSeek-V3 and r1 Glitch tokens (previously) are tokens or strings that trigger strange behavior in LLMs, hinting at oddities in their tokenizers or model weights. Here’s a fun exploration of them across DeepSeek v3 and R1.…
-
Simon Willison’s Weblog: DeepSeek-R1 and exploring DeepSeek-R1-Distill-Llama-8B
Source URL: https://simonwillison.net/2025/Jan/20/deepseek-r1/ Source: Simon Willison’s Weblog Title: DeepSeek-R1 and exploring DeepSeek-R1-Distill-Llama-8B Feedly Summary: DeepSeek are the Chinese AI lab who dropped the best currently available open weights LLM on Christmas day, DeepSeek v3. That model was trained in part using their unreleased R1 “reasoning" model. Today they’ve released R1 itself, along with a whole…
-
Simon Willison’s Weblog: Codestral 25.01
Source URL: https://simonwillison.net/2025/Jan/13/codestral-2501/ Source: Simon Willison’s Weblog Title: Codestral 25.01 Feedly Summary: Codestral 25.01 Brand new code-focused model from Mistral. Unlike the first Codestral this one isn’t (yet) available as open weights. The model has a 256k token context – a new record for Mistral. The new model scored an impressive joint first place with…
-
Hacker News: Notes on the New Deepseek v3
Source URL: https://composio.dev/blog/notes-on-new-deepseek-v3/ Source: Hacker News Title: Notes on the New Deepseek v3 Feedly Summary: Comments AI Summary and Description: Yes **Summary:** The text discusses the release of Deepseek’s v3 model, a 607B mixture-of-experts model that showcases exceptional performance, surpassing both open-source and proprietary competitors at a significantly lower training cost. It highlights the engineering…