Tag: trustworthiness

  • The Register: AI that once called itself MechaHitler will now be available to the US government for $0.42

    Source URL: https://www.theregister.com/2025/09/25/grokai_servces_us_government/ Source: The Register Title: AI that once called itself MechaHitler will now be available to the US government for $0.42 Feedly Summary: Elon Musk’s AI appears to be more ideological than competitors Despite protest letters, concerns that it’s biased and untrustworthy, model tweaks to appease its billionaire boss, and even a past…

  • OpenAI : Measuring the performance of our models on real-world tasks

    Source URL: https://openai.com/index/gdpval Source: OpenAI Title: Measuring the performance of our models on real-world tasks Feedly Summary: OpenAI introduces GDPval-v0, a new evaluation that measures model performance on real-world economically valuable tasks across 44 occupations. AI Summary and Description: Yes Summary: OpenAI’s introduction of GDPval-v0 represents a significant advancement in evaluating AI model performance, particularly…

  • Cloud Blog: Announcing the 2025 DORA Report: State of AI-Assisted Software Development

    Source URL: https://cloud.google.com/blog/products/ai-machine-learning/announcing-the-2025-dora-report/ Source: Cloud Blog Title: Announcing the 2025 DORA Report: State of AI-Assisted Software Development Feedly Summary: Today, we are excited to announce the 2025 DORA Report: State of AI-assisted Software Development. Drawing on insights from over 100 hours of qualitative data and survey responses from nearly 5,000 technology professionals from around the…

  • Slashdot: OpenAI Says Models Programmed To Make Stuff Up Instead of Admitting Ignorance

    Source URL: https://slashdot.org/story/25/09/17/1724241/openai-says-models-programmed-to-make-stuff-up-instead-of-admitting-ignorance?utm_source=rss1.0mainlinkanon&utm_medium=feed Source: Slashdot Title: OpenAI Says Models Programmed To Make Stuff Up Instead of Admitting Ignorance Feedly Summary: AI Summary and Description: Yes Summary: The text discusses OpenAI’s acknowledgment of the issue of “hallucinations” in AI models, specifically how these models frequently yield false outputs due to a training bias that rewards generating…

  • OpenAI : Detecting and reducing scheming in AI models

    Source URL: https://openai.com/index/detecting-and-reducing-scheming-in-ai-models Source: OpenAI Title: Detecting and reducing scheming in AI models Feedly Summary: Apollo Research and OpenAI developed evaluations for hidden misalignment (“scheming”) and found behaviors consistent with scheming in controlled tests across frontier models. The team shared concrete examples and stress tests of an early method to reduce scheming. AI Summary and…

  • The Register: OpenAI says models are programmed to make stuff up instead of admitting ignorance

    Source URL: https://www.theregister.com/2025/09/17/openai_hallucinations_incentives/ Source: The Register Title: OpenAI says models are programmed to make stuff up instead of admitting ignorance Feedly Summary: Even a wrong answer is right some of the time AI models often produce false outputs, or “hallucinations." Now OpenAI has admitted they may result from fundamental mistakes it makes when training its…

  • Simon Willison’s Weblog: Defeating Nondeterminism in LLM Inference

    Source URL: https://simonwillison.net/2025/Sep/11/defeating-nondeterminism/#atom-everything Source: Simon Willison’s Weblog Title: Defeating Nondeterminism in LLM Inference Feedly Summary: Defeating Nondeterminism in LLM Inference A very common question I see about LLMs concerns why they can’t be made to deliver the same response to the same prompt by setting a fixed random number seed. Like many others I had…

  • OpenAI : Why language models hallucinate

    Source URL: https://openai.com/index/why-language-models-hallucinate Source: OpenAI Title: Why language models hallucinate Feedly Summary: OpenAI’s new research explains why language models hallucinate. The findings show how improved evaluations can enhance AI reliability, honesty, and safety. AI Summary and Description: Yes Summary: The text discusses OpenAI’s research on the phenomenon of hallucination in language models, offering insights into…