Tag: single

  • Simon Willison’s Weblog: Constitutional Classifiers: Defending against universal jailbreaks

    Source URL: https://simonwillison.net/2025/Feb/3/constitutional-classifiers/ Source: Simon Willison’s Weblog Title: Constitutional Classifiers: Defending against universal jailbreaks Feedly Summary: Constitutional Classifiers: Defending against universal jailbreaks Interesting new research from Anthropic, resulting in the paper Constitutional Classifiers: Defending against Universal Jailbreaks across Thousands of Hours of Red Teaming. From the paper: In particular, we introduce Constitutional Classifiers, a framework…

  • Cloud Blog: CVE-2023-6080: A Case Study on Third-Party Installer Abuse

    Source URL: https://cloud.google.com/blog/topics/threat-intelligence/cve-2023-6080-third-party-installer-abuse/ Source: Cloud Blog Title: CVE-2023-6080: A Case Study on Third-Party Installer Abuse Feedly Summary: Written By: Jacob Paullus, Daniel McNamara, Jake Rawlins, Steven Karschnia Executive Summary Mandiant exploited flaws in the Microsoft Software Installer (MSI) repair action of Lakeside Software’s SysTrack installer to obtain arbitrary code execution. An attacker with low-privilege access…

  • Simon Willison’s Weblog: OpenAI reasoning models: Advice on prompting

    Source URL: https://simonwillison.net/2025/Feb/2/openai-reasoning-models-advice-on-prompting/ Source: Simon Willison’s Weblog Title: OpenAI reasoning models: Advice on prompting Feedly Summary: OpenAI reasoning models: Advice on prompting OpenAI’s documentation for their o1 and o3 “reasoning models" includes some interesting tips on how to best prompt them: Developer messages are the new system messages: Starting with o1-2024-12-17, reasoning models support developer…

  • Hacker News: Show HN: I built a full mulimodal LLM by merging multiple models into one

    Source URL: https://github.com/JigsawStack/omiai Source: Hacker News Title: Show HN: I built a full mulimodal LLM by merging multiple models into one Feedly Summary: Comments AI Summary and Description: Yes **Short Summary with Insight:** The text presents OmiAI, a highly versatile AI SDK designed specifically for Typescript that streamlines the use of large language models (LLMs).…

  • Wired: DeepSeek’s Safety Guardrails Failed Every Test Researchers Threw at Its AI Chatbot

    Source URL: https://www.wired.com/story/deepseeks-ai-jailbreak-prompt-injection-attacks/ Source: Wired Title: DeepSeek’s Safety Guardrails Failed Every Test Researchers Threw at Its AI Chatbot Feedly Summary: Security researchers tested 50 well-known jailbreaks against DeepSeek’s popular new AI chatbot. It didn’t stop a single one. AI Summary and Description: Yes Summary: The text highlights the ongoing battle between hackers and security researchers…

  • Cloud Blog: Blackwell is here — new A4 VMs powered by NVIDIA B200 now in preview

    Source URL: https://cloud.google.com/blog/products/compute/introducing-a4-vms-powered-by-nvidia-b200-gpu-aka-blackwell/ Source: Cloud Blog Title: Blackwell is here — new A4 VMs powered by NVIDIA B200 now in preview Feedly Summary: Modern AI workloads require powerful accelerators and high-speed interconnects to run sophisticated model architectures on an ever-growing diverse range of model sizes and modalities. In addition to large-scale training, these complex models…

  • Cloud Blog: Announcing the general availability of Spanner Graph

    Source URL: https://cloud.google.com/blog/products/databases/spanner-graph-is-now-ga/ Source: Cloud Blog Title: Announcing the general availability of Spanner Graph Feedly Summary: In today’s complex digital world, building truly intelligent applications requires more than just raw data — you need to understand the intricate relationships within that data. Graph analysis helps reveal these hidden connections, and when combined with techniques like…

  • Simon Willison’s Weblog: How we estimate the risk from prompt injection attacks on AI systems

    Source URL: https://simonwillison.net/2025/Jan/29/prompt-injection-attacks-on-ai-systems/ Source: Simon Willison’s Weblog Title: How we estimate the risk from prompt injection attacks on AI systems Feedly Summary: How we estimate the risk from prompt injection attacks on AI systems The “Agentic AI Security Team" at Google DeepMind share some details on how they are researching indirect prompt injection attacks. They…