Tag: large language model

  • Simon Willison’s Weblog: Defeating Nondeterminism in LLM Inference

    Source URL: https://simonwillison.net/2025/Sep/11/defeating-nondeterminism/#atom-everything Source: Simon Willison’s Weblog Title: Defeating Nondeterminism in LLM Inference Feedly Summary: Defeating Nondeterminism in LLM Inference A very common question I see about LLMs concerns why they can’t be made to deliver the same response to the same prompt by setting a fixed random number seed. Like many others I had…

  • AWS Open Source Blog: Strands Agents and the Model-Driven Approach

    Source URL: https://aws.amazon.com/blogs/opensource/strands-agents-and-the-model-driven-approach/ Source: AWS Open Source Blog Title: Strands Agents and the Model-Driven Approach Feedly Summary: Until recently, building AI agents meant wrestling with complex orchestration frameworks. Developers wrote elaborate state machines, predefined workflows, and extensive error-handling code to guide language models through multi-step tasks. We needed to build elaborate decision trees to handle…

  • Slashdot: HHS Asks All Employees To Start Using ChatGPT

    Source URL: https://yro.slashdot.org/story/25/09/09/2129252/hhs-asks-all-employees-to-start-using-chatgpt?utm_source=rss1.0mainlinkanon&utm_medium=feed Source: Slashdot Title: HHS Asks All Employees To Start Using ChatGPT Feedly Summary: AI Summary and Description: Yes Summary: The email from the Department of Health and Human Services (HHS) announces the rollout of ChatGPT for all employees to improve efficiency. However, there are concerns regarding the management of sensitive information and…

  • Slashdot: Sam Altman Says Bots Are Making Social Media Feel ‘Fake’

    Source URL: https://tech.slashdot.org/story/25/09/09/0048216/sam-altman-says-bots-are-making-social-media-feel-fake?utm_source=rss1.0mainlinkanon&utm_medium=feed Source: Slashdot Title: Sam Altman Says Bots Are Making Social Media Feel ‘Fake’ Feedly Summary: AI Summary and Description: Yes Summary: The text discusses Sam Altman’s observations on the prevalence of bots and AI-generated content on social media platforms, particularly regarding the OpenAI Codex. Altman expresses concern about the authenticity of social…

  • Cloud Blog: Registration now open: Our no-cost, generative AI training and certification program for veterans

    Source URL: https://cloud.google.com/blog/topics/training-certifications/register-for-the-gen-ai-training-and-certification-program-for-veterans/ Source: Cloud Blog Title: Registration now open: Our no-cost, generative AI training and certification program for veterans Feedly Summary: Growing up in a Navy family instilled a strong sense of purpose in me. My father’s remarkable 42 years of naval service not only shaped my values, but inspired me to join the…

  • Slashdot: Microsoft’s Analog Optical Computer Shows AI Promise

    Source URL: https://hardware.slashdot.org/story/25/09/08/0125250/microsofts-analog-optical-computer-shows-ai-promise?utm_source=rss1.0mainlinkanon&utm_medium=feed Source: Slashdot Title: Microsoft’s Analog Optical Computer Shows AI Promise Feedly Summary: AI Summary and Description: Yes Summary: The text discusses a project by Microsoft Research involving an analog optical computer (AOC) designed for AI workloads, significantly enhancing computation speed and energy efficiency compared to traditional GPUs. The initiative offers opportunities for…

  • Simon Willison’s Weblog: Is the LLM response wrong, or have you just failed to iterate it?

    Source URL: https://simonwillison.net/2025/Sep/7/is-the-llm-response-wrong-or-have-you-just-failed-to-iterate-it/#atom-everything Source: Simon Willison’s Weblog Title: Is the LLM response wrong, or have you just failed to iterate it? Feedly Summary: Is the LLM response wrong, or have you just failed to iterate it? More from Mike Caulfield (see also the SIFT method). He starts with a fantastic example of Google’s AI mode…

  • Wired: Psychological Tricks Can Get AI to Break the Rules

    Source URL: https://arstechnica.com/science/2025/09/these-psychological-tricks-can-get-llms-to-respond-to-forbidden-prompts/ Source: Wired Title: Psychological Tricks Can Get AI to Break the Rules Feedly Summary: Researchers convinced large language model chatbots to comply with “forbidden” requests using a variety of conversational tactics. AI Summary and Description: Yes Summary: The text discusses researchers’ exploration of conversational tactics used to manipulate large language model (LLM)…