Tag: Dual Chunk Attention
-
Hacker News: Qwen2.5-7B-Instruct-1M and Qwen2.5-14B-Instruct-1M
Source URL: https://simonwillison.net/2025/Jan/26/qwen25-1m/ Source: Hacker News Title: Qwen2.5-7B-Instruct-1M and Qwen2.5-14B-Instruct-1M Feedly Summary: Comments AI Summary and Description: Yes Summary: The Qwen 2.5 model release from Alibaba introduces a significant advancement in Large Language Model (LLM) capabilities with its ability to process up to 1 million tokens. This increase in input capacity is made possible through…
-
Simon Willison’s Weblog: Qwen2.5-1M: Deploy Your Own Qwen with Context Length up to 1M Tokens
Source URL: https://simonwillison.net/2025/Jan/26/qwen25-1m/ Source: Simon Willison’s Weblog Title: Qwen2.5-1M: Deploy Your Own Qwen with Context Length up to 1M Tokens Feedly Summary: Qwen2.5-1M: Deploy Your Own Qwen with Context Length up to 1M Tokens Very significant new release from Alibaba’s Qwen team. Their openly licensed (sometimes Apache 2, sometimes Qwen license, I’ve had trouble keeping…