Tag: large language model
-
Embrace The Red: Terminal DiLLMa: LLM-powered Apps Can Hijack Your Terminal Via Prompt Injection
Source URL: https://embracethered.com/blog/posts/2024/terminal-dillmas-prompt-injection-ansi-sequences/ Source: Embrace The Red Title: Terminal DiLLMa: LLM-powered Apps Can Hijack Your Terminal Via Prompt Injection Feedly Summary: Last week Leon Derczynski described how LLMs can output ANSI escape codes. These codes, also known as control characters, are interpreted by terminal emulators and modify behavior. This discovery resonates with areas I had…
-
Simon Willison’s Weblog: Roaming RAG – make the model find the answers
Source URL: https://simonwillison.net/2024/Dec/6/roaming-rag/#atom-everything Source: Simon Willison’s Weblog Title: Roaming RAG – make the model find the answers Feedly Summary: Roaming RAG – make the model find the answers Neat new RAG technique (with a snappy name) from John Berryman: The big idea of Roaming RAG is to craft a simple LLM application so that the…
-
Hacker News: AmpereOne: Cores Are the New MHz
Source URL: https://www.jeffgeerling.com/blog/2024/ampereone-cores-are-new-mhz Source: Hacker News Title: AmpereOne: Cores Are the New MHz Feedly Summary: Comments AI Summary and Description: Yes **Summary:** The text provides an in-depth examination of the Supermicro ARS-211ME-FNR server equipped with the 192-core AmpereOne A192-32X CPU, focusing on its design and performance metrics. The analysis highlights how advancements in core technology…
-
Hacker News: Exploring inference memory saturation effect: H100 vs. MI300x
Source URL: https://dstack.ai/blog/h100-mi300x-inference-benchmark/ Source: Hacker News Title: Exploring inference memory saturation effect: H100 vs. MI300x Feedly Summary: Comments AI Summary and Description: Yes **Summary:** The text provides a detailed benchmarking analysis comparing NVIDIA’s H100 GPU and AMD’s MI300x, with a focus on their memory capabilities and implications for LLM (Large Language Model) inference performance. It…
-
Simon Willison’s Weblog: New Pleias 1.0 LLMs trained exclusively on openly licensed data
Source URL: https://simonwillison.net/2024/Dec/5/pleias-llms/#atom-everything Source: Simon Willison’s Weblog Title: New Pleias 1.0 LLMs trained exclusively on openly licensed data Feedly Summary: New Pleias 1.0 LLMs trained exclusively on openly licensed data I wrote about the Common Corpus public domain dataset back in March. Now Pleias, the team behind Common Corpus, have released the first family of…
-
Simon Willison’s Weblog: Claude 3.5 Haiku price drops by 20%
Source URL: https://simonwillison.net/2024/Dec/5/claude-35-haiku-price-drops-by-20/#atom-everything Source: Simon Willison’s Weblog Title: Claude 3.5 Haiku price drops by 20% Feedly Summary: Claude 3.5 Haiku price drops by 20% Buried in this otherwise quite dry post about Anthropic’s ongoing partnership with AWS: To make this model even more accessible for a wide range of use cases, we’re lowering the price…