Tag: sparse attention

  • Simon Willison’s Weblog: Two more Chinese pelicans

    Source URL: https://simonwillison.net/2025/Oct/1/two-pelicans/#atom-everything Source: Simon Willison’s Weblog Title: Two more Chinese pelicans Feedly Summary: Two new models from Chinese AI labs in the past few days. I tried them both out using llm-openrouter: DeepSeek-V3.2-Exp from DeepSeek. Announcement, Tech Report, Hugging Face (690GB, MIT license). As an intermediate step toward our next-generation architecture, V3.2-Exp builds upon…

  • Hacker News: Qwen2.5-1M: Deploy Your Own Qwen with Context Length Up to 1M Tokens

    Source URL: https://qwenlm.github.io/blog/qwen2.5-1m/ Source: Hacker News Title: Qwen2.5-1M: Deploy Your Own Qwen with Context Length Up to 1M Tokens Feedly Summary: Comments AI Summary and Description: Yes Summary: The text reports on the new release of the open-source Qwen2.5-1M models, capable of processing up to one million tokens, significantly improving inference speed and model performance…