Tag: Guardrails

  • The Register: ChatGPT hates LA Chargers fans

    Source URL: https://www.theregister.com/2025/08/27/chatgpt_has_a_problem_with/ Source: The Register Title: ChatGPT hates LA Chargers fans Feedly Summary: Harvard researchers find model guardrails tailor query responses to user’s inferred politics and other affiliations OpenAI’s ChatGPT appears to be more likely to refuse to respond to questions posed by fans of the Los Angeles Chargers football team than to followers…

  • Slashdot: One Long Sentence is All It Takes To Make LLMs Misbehave

    Source URL: https://slashdot.org/story/25/08/27/1756253/one-long-sentence-is-all-it-takes-to-make-llms-misbehave?utm_source=rss1.0mainlinkanon&utm_medium=feed Source: Slashdot Title: One Long Sentence is All It Takes To Make LLMs Misbehave Feedly Summary: AI Summary and Description: Yes Summary: The text discusses a significant security research finding from Palo Alto Networks’ Unit 42 regarding vulnerabilities in large language models (LLMs). The researchers explored methods that allow users to bypass…

  • The Cloudflare Blog: Block unsafe prompts targeting your LLM endpoints with Firewall for AI

    Source URL: https://blog.cloudflare.com/block-unsafe-llm-prompts-with-firewall-for-ai/ Source: The Cloudflare Blog Title: Block unsafe prompts targeting your LLM endpoints with Firewall for AI Feedly Summary: Cloudflare’s AI security suite now includes unsafe content moderation, integrated into the Application Security Suite via Firewall for AI. AI Summary and Description: Yes Summary: The text discusses the launch of Cloudflare’s Firewall for…

  • The Register: One long sentence is all it takes to make LLMs misbehave

    Source URL: https://www.theregister.com/2025/08/26/breaking_llms_for_fun/ Source: The Register Title: One long sentence is all it takes to make LLMs misbehave Feedly Summary: Chatbots ignore their guardrails when your grammar sucks, researchers find Security researchers from Palo Alto Networks’ Unit 42 have discovered the key to getting large language model (LLM) chatbots to ignore their guardrails, and it’s…

  • The Cloudflare Blog: Beyond the ban: A better way to secure generative AI applications

    Source URL: https://blog.cloudflare.com/ai-prompt-protection/ Source: The Cloudflare Blog Title: Beyond the ban: A better way to secure generative AI applications Feedly Summary: Generative AI tools present a trade-off of productivity and data risk. Cloudflare One’s new AI prompt protection feature provides the visibility and control needed to govern these tools, allowing AI Summary and Description: Yes…

  • Docker: A practitioner’s view on how Docker enables security by default and makes developers work better

    Source URL: https://www.docker.com/blog/how-docker-enables-security-by-default/ Source: Docker Title: A practitioner’s view on how Docker enables security by default and makes developers work better Feedly Summary: This blog post was written by Docker Captains, experienced professionals recognized for their expertise with Docker. It shares their firsthand, real-world experiences using Docker in their own work or within the organizations…

  • The Register: LLM chatbots trivial to weaponise for data theft, say boffins

    Source URL: https://www.theregister.com/2025/08/15/llm_chatbots_trivial_to_weaponise/ Source: The Register Title: LLM chatbots trivial to weaponise for data theft, say boffins Feedly Summary: System prompt engineering turns benign AI assistants into ‘investigator’ and ‘detective’ roles that bypass privacy guardrails A team of boffins is warning that AI chatbots built on large language models (LLM) can be tuned into malicious…

  • Cisco Talos Blog: What happened in Vegas (that you actually want to know about)

    Source URL: https://blog.talosintelligence.com/what-happened-in-vegas-that-you-actually-want-to-know-about/ Source: Cisco Talos Blog Title: What happened in Vegas (that you actually want to know about) Feedly Summary: Hazel braves Vegas, overpriced water and the Black Hat maze to bring you Talos’ latest research — including a deep dive into the PS1Bot malware campaign. AI Summary and Description: Yes Summary: This newsletter…