Tag: o1
-
Slashdot: OpenAI’s AI Reasoning Model ‘Thinks’ In Chinese Sometimes, No One Really Knows Why
Source URL: https://slashdot.org/story/25/01/14/239246/openais-ai-reasoning-model-thinks-in-chinese-sometimes-no-one-really-knows-why?utm_source=rss1.0mainlinkanon&utm_medium=feed Source: Slashdot Title: OpenAI’s AI Reasoning Model ‘Thinks’ In Chinese Sometimes, No One Really Knows Why Feedly Summary: AI Summary and Description: Yes Summary: The behavior exhibited by OpenAI’s reasoning AI model, o1, which seemingly “thinks” in multiple languages regardless of the input language, has raised questions within the AI community. Experts…
-
Simon Willison’s Weblog: Codestral 25.01
Source URL: https://simonwillison.net/2025/Jan/13/codestral-2501/ Source: Simon Willison’s Weblog Title: Codestral 25.01 Feedly Summary: Codestral 25.01 Brand new code-focused model from Mistral. Unlike the first Codestral this one isn’t (yet) available as open weights. The model has a 256k token context – a new record for Mistral. The new model scored an impressive joint first place with…
-
Simon Willison’s Weblog: Quoting Ben Hylak
Source URL: https://simonwillison.net/2025/Jan/12/ben-hylak/#atom-everything Source: Simon Willison’s Weblog Title: Quoting Ben Hylak Feedly Summary: I was using o1 like a chat model — but o1 is not a chat model. If o1 is not a chat model — what is it? I think of it like a “report generator.” If you give it enough context, and…
-
Hacker News: Contemplative LLMs
Source URL: https://maharshi.bearblog.dev/contemplative-llms-prompt/ Source: Hacker News Title: Contemplative LLMs Feedly Summary: Comments AI Summary and Description: Yes **Short Summary with Insight:** The text discusses the novel approach of prompting Large Language Models (LLMs) to engage in a contemplation phase before generating answers. By mimicking a reasoning process, which encourages exploration and questioning assumptions, this method…
-
OpenAI : Deliberative alignment: reasoning enables safer language models
Source URL: https://openai.com/index/deliberative-alignment Source: OpenAI Title: Deliberative alignment: reasoning enables safer language models Feedly Summary: Deliberative alignment: reasoning enables safer language models Introducing our new alignment strategy for o1 models, which are directly taught safety specifications and how to reason over them. AI Summary and Description: Yes Summary: The text discusses a new alignment strategy…
-
The Register: Even at $200/mo, Altman admits ChatGPT Pro struggles to turn a profit
Source URL: https://www.theregister.com/2025/01/06/altman_gpt_profits/ Source: The Register Title: Even at $200/mo, Altman admits ChatGPT Pro struggles to turn a profit Feedly Summary: But don’t worry, he’s ‘figured out’ AGI comment Even at $200 a month for ChatGPT Pro, the service is struggling to turn a profit, OpenAI CEO Sam Altman lamented on the platform formerly known…
-
Simon Willison’s Weblog: Quoting François Chollet
Source URL: https://simonwillison.net/2025/Jan/6/francois-chollet/#atom-everything Source: Simon Willison’s Weblog Title: Quoting François Chollet Feedly Summary: I don’t think people really appreciate how simple ARC-AGI-1 was, and what solving it really means. It was designed as the simplest, most basic assessment of fluid intelligence possible. Failure to pass signifies a near-total inability to adapt or problem-solve in unfamiliar…
-
Simon Willison’s Weblog: Using LLMs and Cursor to become a finisher
Source URL: https://simonwillison.net/2025/Jan/4/llms-and-cursor/#atom-everything Source: Simon Willison’s Weblog Title: Using LLMs and Cursor to become a finisher Feedly Summary: Using LLMs and Cursor to become a finisher Zohaib Rauf describes a pattern I’ve seen quite a few examples of now: engineers who moved into management but now find themselves able to ship working code again (at…
-
Hacker News: A path to O1 open source
Source URL: https://arxiv.org/abs/2412.14135 Source: Hacker News Title: A path to O1 open source Feedly Summary: Comments AI Summary and Description: Yes Summary: The text discusses advancements in artificial intelligence, particularly focusing on the reinforcement learning approach to reproduce OpenAI’s o1 model. It highlights key components like policy initialization, reward design, search, and learning that contribute…