Tag: jailbreak

  • Hacker News: Constitutional Classifiers: Defending against universal jailbreaks

    Source URL: https://www.anthropic.com/research/constitutional-classifiers Source: Hacker News Title: Constitutional Classifiers: Defending against universal jailbreaks Feedly Summary: Comments AI Summary and Description: Yes Summary: The text discusses a novel approach by the Anthropic Safeguards Research Team to defend AI models against jailbreaks through the use of Constitutional Classifiers. This system demonstrates robustness against various jailbreak techniques while…

  • Hacker News: Notes on OpenAI O3-Mini

    Source URL: https://simonwillison.net/2025/Jan/31/o3-mini/ Source: Hacker News Title: Notes on OpenAI O3-Mini Feedly Summary: Comments AI Summary and Description: Yes Summary: The announcement of OpenAI’s o3-mini model marks a significant development in the landscape of large language models (LLMs). With enhanced performance on specific benchmarks and user functionalities that include internet search capabilities, o3-mini aims to…

  • Simon Willison’s Weblog: OpenAI o3-mini, now available in LLM

    Source URL: https://simonwillison.net/2025/Jan/31/o3-mini/#atom-everything Source: Simon Willison’s Weblog Title: OpenAI o3-mini, now available in LLM Feedly Summary: o3-mini is out today. As with other o-series models it’s a slightly difficult one to evaluate – we now need to decide if a prompt is best run using GPT-4o, o1, o3-mini or (if we have access) o1 Pro.…

  • Wired: DeepSeek’s Safety Guardrails Failed Every Test Researchers Threw at Its AI Chatbot

    Source URL: https://www.wired.com/story/deepseeks-ai-jailbreak-prompt-injection-attacks/ Source: Wired Title: DeepSeek’s Safety Guardrails Failed Every Test Researchers Threw at Its AI Chatbot Feedly Summary: Security researchers tested 50 well-known jailbreaks against DeepSeek’s popular new AI chatbot. It didn’t stop a single one. AI Summary and Description: Yes Summary: The text highlights the ongoing battle between hackers and security researchers…

  • Hacker News: O3-mini System Card [pdf]

    Source URL: https://cdn.openai.com/o3-mini-system-card.pdf Source: Hacker News Title: O3-mini System Card [pdf] Feedly Summary: Comments AI Summary and Description: Yes **Summary:** The OpenAI o3-mini System Card details the advanced capabilities, safety evaluations, and risk classifications of the OpenAI o3-mini model. This document is particularly pertinent for professionals in AI security, as it outlines significant safety measures…

  • The Register: Google to Iran: Yes, we see you using Gemini for phishing and scripting. We’re onto you

    Source URL: https://www.theregister.com/2025/01/31/state_spies_google_gemini/ Source: The Register Title: Google to Iran: Yes, we see you using Gemini for phishing and scripting. We’re onto you Feedly Summary: And you, China, Russia, North Korea … Guardrails block malware generation Google says it’s spotted Chinese, Russian, Iranian, and North Korean government agents using its Gemini AI for nefarious purposes,…

  • Unit 42: Recent Jailbreaks Demonstrate Emerging Threat to DeepSeek

    Source URL: https://unit42.paloaltonetworks.com/?p=138180 Source: Unit 42 Title: Recent Jailbreaks Demonstrate Emerging Threat to DeepSeek Feedly Summary: Evaluation of three jailbreaking techniques on DeepSeek shows risks of generating prohibited content. The post Recent Jailbreaks Demonstrate Emerging Threat to DeepSeek appeared first on Unit 42. AI Summary and Description: Yes Summary: The text outlines the research conducted…

  • Simon Willison’s Weblog: How we estimate the risk from prompt injection attacks on AI systems

    Source URL: https://simonwillison.net/2025/Jan/29/prompt-injection-attacks-on-ai-systems/ Source: Simon Willison’s Weblog Title: How we estimate the risk from prompt injection attacks on AI systems Feedly Summary: How we estimate the risk from prompt injection attacks on AI systems The “Agentic AI Security Team" at Google DeepMind share some details on how they are researching indirect prompt injection attacks. They…

  • Hacker News: 1,156 Questions Censored by DeepSeek

    Source URL: https://www.promptfoo.dev/blog/deepseek-censorship/ Source: Hacker News Title: 1,156 Questions Censored by DeepSeek Feedly Summary: Comments AI Summary and Description: Yes **Summary**: The text discusses the DeepSeek-R1 model, highlighting its prominence and the associated concerns regarding censorship driven by CCP policies. It emphasizes the model’s high refusal rate on sensitive topics in China, the methods to…