Tag: jailbreak
-
OpenAI : GPT-5 bio bug bounty call
Source URL: https://openai.com/gpt-5-bio-bug-bounty Source: OpenAI Title: GPT-5 bio bug bounty call Feedly Summary: OpenAI invites researchers to its Bio Bug Bounty. Test GPT-5’s safety with a universal jailbreak prompt and win up to $25,000. AI Summary and Description: Yes Summary: OpenAI’s initiative invites researchers to participate in its Bio Bug Bounty program, focusing on testing…
-
The Register: LegalPwn: Tricking LLMs by burying badness in lawyerly fine print
Source URL: https://www.theregister.com/2025/09/01/legalpwn_ai_jailbreak/ Source: The Register Title: LegalPwn: Tricking LLMs by burying badness in lawyerly fine print Feedly Summary: Trust and believe – AI models trained to see ‘legal’ doc as super legit Researchers at security firm Pangea have discovered yet another way to trivially trick large language models (LLMs) into ignoring their guardrails. Stick…
-
OpenAI : OpenAI and Anthropic share findings from a joint safety evaluation
Source URL: https://openai.com/index/openai-anthropic-safety-evaluation Source: OpenAI Title: OpenAI and Anthropic share findings from a joint safety evaluation Feedly Summary: OpenAI and Anthropic share findings from a first-of-its-kind joint safety evaluation, testing each other’s models for misalignment, instruction following, hallucinations, jailbreaking, and more—highlighting progress, challenges, and the value of cross-lab collaboration. AI Summary and Description: Yes Summary:…
-
Slashdot: Red Teams Jailbreak GPT-5 With Ease, Warn It’s ‘Nearly Unusable’ For Enterprise
Source URL: https://it.slashdot.org/story/25/08/08/2113251/red-teams-jailbreak-gpt-5-with-ease-warn-its-nearly-unusable-for-enterprise?utm_source=rss1.0mainlinkanon&utm_medium=feed Source: Slashdot Title: Red Teams Jailbreak GPT-5 With Ease, Warn It’s ‘Nearly Unusable’ For Enterprise Feedly Summary: AI Summary and Description: Yes Summary: The text highlights significant security vulnerabilities in the newly released GPT-5 model, noting that it was easily jailbroken within a short timeframe. The results from different red teaming efforts…
-
OpenAI : Agent bio bug bounty call
Source URL: https://openai.com/bio-bug-bounty Source: OpenAI Title: Agent bio bug bounty call Feedly Summary: OpenAI invites researchers to its Bio Bug Bounty. Test the ChatGPT agent’s safety with a universal jailbreak prompt and win up to $25,000. AI Summary and Description: Yes Summary: The text highlights OpenAI’s Bio Bug Bounty initiative, which invites researchers to test…
-
The Register: How to trick ChatGPT into revealing Windows keys? I give up
Source URL: https://www.theregister.com/2025/07/09/chatgpt_jailbreak_windows_keys/ Source: The Register Title: How to trick ChatGPT into revealing Windows keys? I give up Feedly Summary: No, really, those are the magic words A clever AI bug hunter found a way to trick ChatGPT into disclosing Windows product keys, including at least one owned by Wells Fargo bank, by inviting the…
-
Cisco Talos Blog: Cybercriminal abuse of large language models
Source URL: https://blog.talosintelligence.com/cybercriminal-abuse-of-large-language-models/ Source: Cisco Talos Blog Title: Cybercriminal abuse of large language models Feedly Summary: Cybercriminals are increasingly gravitating towards uncensored LLMs, cybercriminal-designed LLMs and jailbreaking legitimate LLMs. AI Summary and Description: Yes **Summary:** The provided text discusses how cybercriminals exploit artificial intelligence technologies, particularly large language models (LLMs), to enhance their criminal activities.…