Tag: mathematics
-
Hacker News: SmolLM2
Source URL: https://simonwillison.net/2024/Nov/2/smollm2/ Source: Hacker News Title: SmolLM2 Feedly Summary: Comments AI Summary and Description: Yes Summary: The text introduces SmolLM2, a new family of compact language models from Hugging Face, designed for lightweight on-device operations. The models, which range from 135M to 1.7B parameters, were trained on 11 trillion tokens across diverse datasets, showcasing…
-
Simon Willison’s Weblog: SmolLM2
Source URL: https://simonwillison.net/2024/Nov/2/smollm2/#atom-everything Source: Simon Willison’s Weblog Title: SmolLM2 Feedly Summary: SmolLM2 New from Loubna Ben Allal and her research team at Hugging Face: SmolLM2 is a family of compact language models available in three size: 135M, 360M, and 1.7B parameters. They are capable of solving a wide range of tasks while being lightweight enough…
-
Hacker News: DeepSeek: Advancing theorem proving in LLMs through large-scale synthetic data
Source URL: https://arxiv.org/abs/2405.14333 Source: Hacker News Title: DeepSeek: Advancing theorem proving in LLMs through large-scale synthetic data Feedly Summary: Comments AI Summary and Description: Yes Summary: The paper introduces DeepSeek-Prover, an innovative approach that leverages large-scale synthetic data to improve the capabilities of large language models (LLMs) in formal theorem proving. It highlights the challenges…
-
New York Times – Artificial Intelligence : Can Math Help AI Chatbots Stop Making Stuff Up?
Source URL: https://www.nytimes.com/2024/09/23/technology/ai-chatbots-chatgpt-math.html Source: New York Times – Artificial Intelligence Title: Can Math Help AI Chatbots Stop Making Stuff Up? Feedly Summary: Chatbots like ChatGPT get stuff wrong. But researchers are building new A.I. systems that can verify their own math — and maybe more. AI Summary and Description: Yes Summary: The text discusses advancements…
-
Simon Willison’s Weblog: Quoting Terrence Tao
Source URL: https://simonwillison.net/2024/Sep/15/terrence-tao/#atom-everything Source: Simon Willison’s Weblog Title: Quoting Terrence Tao Feedly Summary: [… OpenAI’s o1] could work its way to a correct (and well-written) solution if provided a lot of hints and prodding, but did not generate the key conceptual ideas on its own, and did make some non-trivial mistakes. The experience seemed roughly…
-
Simon Willison’s Weblog: Quoting Jason Wei (OpenAI)
Source URL: https://simonwillison.net/2024/Sep/12/jason-wei-openai/#atom-everything Source: Simon Willison’s Weblog Title: Quoting Jason Wei (OpenAI) Feedly Summary: o1-mini is the most surprising research result I’ve seen in the past year Obviously I cannot spill the secret, but a small model getting >60% on AIME math competition is so good that it’s hard to believe— Jason Wei (OpenAI) Tags:…