Tag: jailbreak

  • The Register: Google declares AI bug hunting season open, sets a $30K max reward

    Source URL: https://www.theregister.com/2025/10/07/google_ai_bug_bounty/ Source: The Register Title: Google declares AI bug hunting season open, sets a $30K max reward Feedly Summary: Jailbreaks, direct prompt injection not allowed Google on Monday rolled out a new AI Vulnerability Reward Program to encourage researchers to find and report flaws in its AI systems, with rewards of up to…

  • OpenAI : GPT-5 bio bug bounty call

    Source URL: https://openai.com/gpt-5-bio-bug-bounty Source: OpenAI Title: GPT-5 bio bug bounty call Feedly Summary: OpenAI invites researchers to its Bio Bug Bounty. Test GPT-5’s safety with a universal jailbreak prompt and win up to $25,000. AI Summary and Description: Yes Summary: OpenAI’s initiative invites researchers to participate in its Bio Bug Bounty program, focusing on testing…

  • The Register: LegalPwn: Tricking LLMs by burying badness in lawyerly fine print

    Source URL: https://www.theregister.com/2025/09/01/legalpwn_ai_jailbreak/ Source: The Register Title: LegalPwn: Tricking LLMs by burying badness in lawyerly fine print Feedly Summary: Trust and believe – AI models trained to see ‘legal’ doc as super legit Researchers at security firm Pangea have discovered yet another way to trivially trick large language models (LLMs) into ignoring their guardrails. Stick…

  • OpenAI : OpenAI and Anthropic share findings from a joint safety evaluation

    Source URL: https://openai.com/index/openai-anthropic-safety-evaluation Source: OpenAI Title: OpenAI and Anthropic share findings from a joint safety evaluation Feedly Summary: OpenAI and Anthropic share findings from a first-of-its-kind joint safety evaluation, testing each other’s models for misalignment, instruction following, hallucinations, jailbreaking, and more—highlighting progress, challenges, and the value of cross-lab collaboration. AI Summary and Description: Yes Summary:…

  • Slashdot: Red Teams Jailbreak GPT-5 With Ease, Warn It’s ‘Nearly Unusable’ For Enterprise

    Source URL: https://it.slashdot.org/story/25/08/08/2113251/red-teams-jailbreak-gpt-5-with-ease-warn-its-nearly-unusable-for-enterprise?utm_source=rss1.0mainlinkanon&utm_medium=feed Source: Slashdot Title: Red Teams Jailbreak GPT-5 With Ease, Warn It’s ‘Nearly Unusable’ For Enterprise Feedly Summary: AI Summary and Description: Yes Summary: The text highlights significant security vulnerabilities in the newly released GPT-5 model, noting that it was easily jailbroken within a short timeframe. The results from different red teaming efforts…

  • OpenAI : Agent bio bug bounty call

    Source URL: https://openai.com/bio-bug-bounty Source: OpenAI Title: Agent bio bug bounty call Feedly Summary: OpenAI invites researchers to its Bio Bug Bounty. Test the ChatGPT agent’s safety with a universal jailbreak prompt and win up to $25,000. AI Summary and Description: Yes Summary: The text highlights OpenAI’s Bio Bug Bounty initiative, which invites researchers to test…

  • The Register: How to trick ChatGPT into revealing Windows keys? I give up

    Source URL: https://www.theregister.com/2025/07/09/chatgpt_jailbreak_windows_keys/ Source: The Register Title: How to trick ChatGPT into revealing Windows keys? I give up Feedly Summary: No, really, those are the magic words A clever AI bug hunter found a way to trick ChatGPT into disclosing Windows product keys, including at least one owned by Wells Fargo bank, by inviting the…