Tag: Qwen

  • Simon Willison’s Weblog: What’s new in the world of LLMs, for NICAR 2025

    Source URL: https://simonwillison.net/2025/Mar/8/nicar-llms/ Source: Simon Willison’s Weblog Title: What’s new in the world of LLMs, for NICAR 2025 Feedly Summary: I presented two sessions at the NICAR 2025 data journalism conference this year. The first was this one based on my review of LLMs in 2024, extended by several months to cover everything that’s happened…

  • Hacker News: Ladder: Self-Improving LLMs Through Recursive Problem Decomposition

    Source URL: https://arxiv.org/abs/2503.00735 Source: Hacker News Title: Ladder: Self-Improving LLMs Through Recursive Problem Decomposition Feedly Summary: Comments AI Summary and Description: Yes Summary: The paper introduces LADDER, a novel framework for enhancing the problem-solving capabilities of Large Language Models (LLMs) through a self-guided learning approach. By recursively generating simpler problem variants, LADDER enables models to…

  • Hacker News: Using GRPO to Beat o1, o3-mini and R1 at "Temporal Clue"

    Source URL: https://openpipe.ai/blog/using-grpo-to-beat-o1-o3-mini-and-r1-on-temporal-clue Source: Hacker News Title: Using GRPO to Beat o1, o3-mini and R1 at "Temporal Clue" Feedly Summary: Comments AI Summary and Description: Yes Short Summary with Insight: The provided text explores the application of reinforcement learning to enhance the deductive reasoning capabilities of smaller, open-weight models in AI. Specifically, it focuses on…

  • Hacker News: AMD Announces "Instella" Open-Source 3B Language Models

    Source URL: https://www.phoronix.com/news/AMD-Intella-Open-Source-LM Source: Hacker News Title: AMD Announces "Instella" Open-Source 3B Language Models Feedly Summary: Comments AI Summary and Description: Yes Summary: AMD has announced the open-sourcing of its Instella language models, a significant advancement in the AI domain that promotes transparency, collaboration, and innovation. These models, based on the high-performance MI300X GPUs, aim…

  • Simon Willison’s Weblog: QwQ-32B: Embracing the Power of Reinforcement Learning

    Source URL: https://simonwillison.net/2025/Mar/5/qwq-32b/#atom-everything Source: Simon Willison’s Weblog Title: QwQ-32B: Embracing the Power of Reinforcement Learning Feedly Summary: QwQ-32B: Embracing the Power of Reinforcement Learning New Apache 2 licensed reasoning model from Qwen: We are excited to introduce QwQ-32B, a model with 32 billion parameters that achieves performance comparable to DeepSeek-R1, which boasts 671 billion parameters…

  • Hacker News: QwQ-32B: Embracing the Power of Reinforcement Learning

    Source URL: https://qwenlm.github.io/blog/qwq-32b/ Source: Hacker News Title: QwQ-32B: Embracing the Power of Reinforcement Learning Feedly Summary: Comments AI Summary and Description: Yes Summary: The text discusses the advancements in Reinforcement Learning (RL) as applied to large language models, particularly highlighting the launch of the QwQ-32B model. It emphasizes the model’s performance enhancements through RL and…

  • Schneier on Security: “Emergent Misalignment” in LLMs

    Source URL: https://www.schneier.com/blog/archives/2025/02/emergent-misalignment-in-llms.html Source: Schneier on Security Title: “Emergent Misalignment” in LLMs Feedly Summary: Interesting research: “Emergent Misalignment: Narrow finetuning can produce broadly misaligned LLMs“: Abstract: We present a surprising result regarding LLMs and alignment. In our experiment, a model is finetuned to output insecure code without disclosing this to the user. The resulting model…

  • The Register: Does terrible code drive you mad? Wait until you see what it does to OpenAI’s GPT-4o

    Source URL: https://www.theregister.com/2025/02/27/llm_emergent_misalignment_study/ Source: The Register Title: Does terrible code drive you mad? Wait until you see what it does to OpenAI’s GPT-4o Feedly Summary: Model was fine-tuned to write vulnerable software – then suggested enslaving humanity Computer scientists have found that fine-tuning notionally safe large language models to do one thing badly can negatively…

  • Simon Willison’s Weblog: olmOCR

    Source URL: https://simonwillison.net/2025/Feb/26/olmocr/#atom-everything Source: Simon Willison’s Weblog Title: olmOCR Feedly Summary: olmOCR New from Ai2 – olmOCR is “an open-source tool designed for high-throughput conversion of PDFs and other documents into plain text while preserving natural reading order". At its core is allenai/olmOCR-7B-0225-preview, a Qwen2-VL-7B-Instruct variant trained on ~250,000 pages of diverse PDF content (both…

  • Simon Willison’s Weblog: Quoting Emergent Misalignment: Narrow finetuning can produce broadly misaligned LLMs

    Source URL: https://simonwillison.net/2025/Feb/25/emergent-misalignment/ Source: Simon Willison’s Weblog Title: Quoting Emergent Misalignment: Narrow finetuning can produce broadly misaligned LLMs Feedly Summary: In our experiment, a model is finetuned to output insecure code without disclosing this to the user. The resulting model acts misaligned on a broad range of prompts that are unrelated to coding: it asserts…