Tag: Apache 2

  • The Register: OpenAI makes good on its name, launches first open weights language models since GPT-2

    Source URL: https://www.theregister.com/2025/08/05/openai_open_gpt/ Source: The Register Title: OpenAI makes good on its name, launches first open weights language models since GPT-2 Feedly Summary: GPT-OSS now available in 120 and 20 billion parameter sizes under Apache 2.0 license OpenAI released its first open weights language models since GPT-2 on Tuesday with the debut of GPT-OSS.… AI…

  • Simon Willison’s Weblog: OpenAI’s new open weight (Apache 2) models are really good

    Source URL: https://simonwillison.net/2025/Aug/5/gpt-oss/ Source: Simon Willison’s Weblog Title: OpenAI’s new open weight (Apache 2) models are really good Feedly Summary: The long promised OpenAI open weight models are here, and they are very impressive. They’re available under proper open source licenses – Apache 2.0 – and come in two sizes, 120B and 20B. OpenAI’s own…

  • OpenAI : Introducing gpt-oss

    Source URL: https://openai.com/index/introducing-gpt-oss Source: OpenAI Title: Introducing gpt-oss Feedly Summary: We’re releasing gpt-oss-120b and gpt-oss-20b—two state-of-the-art open-weight language models that deliver strong real-world performance at low cost. Available under the flexible Apache 2.0 license, these models outperform similarly sized open models on reasoning tasks, demonstrate strong tool use capabilities, and are optimized for efficient deployment…

  • OpenAI : gpt-oss-120b & gpt-oss-20b Model Card

    Source URL: https://openai.com/index/gpt-oss-model-card Source: OpenAI Title: gpt-oss-120b & gpt-oss-20b Model Card Feedly Summary: We introduce gpt-oss-120b and gpt-oss-20b, two open-weight reasoning models available under the Apache 2.0 license and our gpt-oss usage policy. AI Summary and Description: Yes Summary: The introduction of gpt-oss-120b and gpt-oss-20b highlights the development of open-weight reasoning models, which reflects significant…

  • Simon Willison’s Weblog: XBai o4

    Source URL: https://simonwillison.net/2025/Aug/3/xbai-o4/#atom-everything Source: Simon Willison’s Weblog Title: XBai o4 Feedly Summary: XBai o4 Yet another open source (Apache 2.0) LLM from a Chinese AI lab. This model card claims: XBai o4 excels in complex reasoning capabilities and has now completely surpassed OpenAI-o3-mini in Medium mode. This a 32.8 billion parameter model released by MetaStone…

  • Simon Willison’s Weblog: The best available open weight LLMs now come from China

    Source URL: https://simonwillison.net/2025/Jul/30/chinese-models/ Source: Simon Willison’s Weblog Title: The best available open weight LLMs now come from China Feedly Summary: Something that has become undeniable this month is that the best available open weight models now come from the Chinese AI labs. I continue to have a lot of love for Mistral, Gemma and Llama…

  • Simon Willison’s Weblog: Qwen3-Coder: Agentic Coding in the World

    Source URL: https://simonwillison.net/2025/Jul/22/qwen3-coder/ Source: Simon Willison’s Weblog Title: Qwen3-Coder: Agentic Coding in the World Feedly Summary: Qwen3-Coder: Agentic Coding in the World It turns out that as I was typing up my notes on Qwen3-235B-A22B-Instruct-2507 the Qwen team were unleashing something much bigger: Today, we’re announcing Qwen3-Coder, our most agentic code model to date. Qwen3-Coder…

  • Simon Willison’s Weblog: Qwen/Qwen3-235B-A22B-Instruct-2507

    Source URL: https://simonwillison.net/2025/Jul/22/qwen3-235b-a22b-instruct-2507/#atom-everything Source: Simon Willison’s Weblog Title: Qwen/Qwen3-235B-A22B-Instruct-2507 Feedly Summary: Qwen/Qwen3-235B-A22B-Instruct-2507 Significant new model release from Qwen, published yesterday without much fanfare. This is a follow-up to their April release of the full Qwen 3 model family, which included a Qwen3-235B-A22B model which could handle both reasoning and non-reasoning prompts (via a /no_think toggle).…

  • Simon Willison’s Weblog: Voxtral

    Source URL: https://simonwillison.net/2025/Jul/16/voxtral/#atom-everything Source: Simon Willison’s Weblog Title: Voxtral Feedly Summary: Voxtral Mistral released their first audio-input models yesterday: Voxtral Small and Voxtral Mini. These state‑of‑the‑art speech understanding models are available in two sizes—a 24B variant for production-scale applications and a 3B variant for local and edge deployments. Both versions are released under the Apache…