Tag: safety measures
-
The Register: Google Gemini tells grad student to ‘please die’ after helping with his homework
Source URL: https://www.theregister.com/2024/11/15/google_gemini_prompt_bad_response/ Source: The Register Title: Google Gemini tells grad student to ‘please die’ after helping with his homework Feedly Summary: First true sign of AGI – blowing a fuse with a frustrating user? When you’re trying to get homework help from an AI model like Google Gemini, the last thing you’d expect is…
-
The Register: Want advice from UK government website about tax ‘n’ stuff? Talk to the chatbot
Source URL: https://www.theregister.com/2024/11/06/uk_government_website_chatbot/ Source: The Register Title: Want advice from UK government website about tax ‘n’ stuff? Talk to the chatbot Feedly Summary: Need to know how to set up a business? There’s an (experimental) AI for that From the department of “this will go well" comes confirmation UK government is trialling an experimental chatbot…
-
The Register: Voice-enabled AI agents can automate everything, even your phone scams
Source URL: https://www.theregister.com/2024/10/24/openai_realtime_api_phone_scam/ Source: The Register Title: Voice-enabled AI agents can automate everything, even your phone scams Feedly Summary: All for the low, low price of a mere dollar Scammers, rejoice. OpenAI’s real-time voice API can be used to build AI agents capable of conducting successful phone call scams for less than a dollar.… AI…
-
The Register: Anthropic’s latest Claude model can interact with computers – what could go wrong?
Source URL: https://www.theregister.com/2024/10/24/anthropic_claude_model_can_use_computers/ Source: The Register Title: Anthropic’s latest Claude model can interact with computers – what could go wrong? Feedly Summary: For starters, it could launch a prompt injection attack on itself… The latest version of AI startup Anthropic’s Claude 3.5 Sonnet model can use computers – and the developer makes it sound like…
-
METR Blog – METR: Common Elements of Frontier AI Safety Policies
Source URL: https://metr.org/blog/2024-08-29-common-elements-of-frontier-ai-safety-policies/ Source: METR Blog – METR Title: Common Elements of Frontier AI Safety Policies Feedly Summary: AI Summary and Description: Yes Summary: The text discusses the Frontier AI Safety Commitments made by sixteen developers of large foundation models at the AI Seoul Summit, which focus on risk evaluation and mitigation strategies to ensure…
-
Hacker News: Computer use, a new Claude 3.5 Sonnet, and Claude 3.5 Haiku
Source URL: https://www.anthropic.com/news/3-5-models-and-computer-use Source: Hacker News Title: Computer use, a new Claude 3.5 Sonnet, and Claude 3.5 Haiku Feedly Summary: Comments AI Summary and Description: Yes Summary: The announcement introduces upgrades to the Claude AI models, particularly highlighting advancements in coding capabilities and the new feature of “computer use,” allowing the AI to interact with…
-
Hacker News: Announcing Our Updated Responsible Scaling Policy
Source URL: https://www.anthropic.com/news/announcing-our-updated-responsible-scaling-policy Source: Hacker News Title: Announcing Our Updated Responsible Scaling Policy Feedly Summary: Comments AI Summary and Description: Yes Summary: The text discusses an important update to the Responsible Scaling Policy (RSP) by Anthropic, aimed at mitigating risks associated with frontier AI systems. The update introduces a robust framework for evaluating AI capabilities…
-
The Register: Anthropic’s Claude vulnerable to ’emotional manipulation’
Source URL: https://www.theregister.com/2024/10/12/anthropics_claude_vulnerable_to_emotional/ Source: The Register Title: Anthropic’s Claude vulnerable to ’emotional manipulation’ Feedly Summary: AI model safety only goes so far Anthropic’s Claude 3.5 Sonnet, despite its reputation as one of the better behaved generative AI models, can still be convinced to emit racist hate speech and malware.… AI Summary and Description: Yes Summary:…
-
Hacker News: Geoffrey Hinton says he’s proud Ilya Sutskever ‘fired Sam Altman’
Source URL: https://techcrunch.com/2024/10/09/after-winning-nobel-for-foundational-ai-work-geoffrey-hinton-says-hes-proud-ilya-sutskever-fired-sam-altman/ Source: Hacker News Title: Geoffrey Hinton says he’s proud Ilya Sutskever ‘fired Sam Altman’ Feedly Summary: Comments AI Summary and Description: Yes Summary: The text discusses Geoffrey Hinton’s acceptance of a Nobel Prize for his contributions to artificial neural networks, highlighting a significant event in the AI community involving the controversial firing…