Tag: position embeddings
-
Simon Willison’s Weblog: deepseek-ai/DeepSeek-V3-Base
Source URL: https://simonwillison.net/2024/Dec/25/deepseek-v3/#atom-everything Source: Simon Willison’s Weblog Title: deepseek-ai/DeepSeek-V3-Base Feedly Summary: deepseek-ai/DeepSeek-V3-Base No model card or announcement yet, but this new model release from Chinese AI lab DeepSeek (an arm of Chinese hedge fund High-Flyer) looks very significant. It’s a huge model – 685B parameters, 687.9 GB on disk (TIL how to size a git-lfs…
-
Simon Willison’s Weblog: Quoting François Chollet
Source URL: https://simonwillison.net/2024/Oct/16/francois-chollet/ Source: Simon Willison’s Weblog Title: Quoting François Chollet Feedly Summary: A common misconception about Transformers is to believe that they’re a sequence-processing architecture. They’re not. They’re a set-processing architecture. Transformers are 100% order-agnostic (which was the big innovation compared to RNNs, back in late 2016 — you compute the full matrix of…