Tag: large language model
-
Simon Willison’s Weblog: Quoting Menlo Ventures
Source URL: https://simonwillison.net/2024/Nov/29/menlo-ventures/#atom-everything Source: Simon Willison’s Weblog Title: Quoting Menlo Ventures Feedly Summary: Among closed-source models, OpenAI’s early mover advantage has eroded somewhat, with enterprise market share dropping from 50% to 34%. The primary beneficiary has been Anthropic,* which doubled its enterprise presence from 12% to 24% as some enterprises switched from GPT-4 to Claude…
-
Simon Willison’s Weblog: Quoting Andrej Karpathy
Source URL: https://simonwillison.net/2024/Nov/29/andrej-karpathy/#atom-everything Source: Simon Willison’s Weblog Title: Quoting Andrej Karpathy Feedly Summary: People have too inflated sense of what it means to “ask an AI" about something. The AI are language models trained basically by imitation on data from human labelers. Instead of the mysticism of "asking an AI", think of it more as…
-
Simon Willison’s Weblog: LLM Flowbreaking
Source URL: https://simonwillison.net/2024/Nov/29/llm-flowbreaking/#atom-everything Source: Simon Willison’s Weblog Title: LLM Flowbreaking Feedly Summary: LLM Flowbreaking Gadi Evron from Knostic: We propose that LLM Flowbreaking, following jailbreaking and prompt injection, joins as the third on the growing list of LLM attack types. Flowbreaking is less about whether prompt or response guardrails can be bypassed, and more about…
-
Hacker News: An Intuitive Explanation of Sparse Autoencoders for LLM Interpretability
Source URL: https://adamkarvonen.github.io/machine_learning/2024/06/11/sae-intuitions.html Source: Hacker News Title: An Intuitive Explanation of Sparse Autoencoders for LLM Interpretability Feedly Summary: Comments AI Summary and Description: Yes **Summary**: The text discusses Sparse Autoencoders (SAEs) and their significance in interpreting machine learning models, particularly large language models (LLMs). It explains how SAEs can provide insights into the functioning of…
-
Simon Willison’s Weblog: QwQ: Reflect Deeply on the Boundaries of the Unknown
Source URL: https://simonwillison.net/2024/Nov/27/qwq/#atom-everything Source: Simon Willison’s Weblog Title: QwQ: Reflect Deeply on the Boundaries of the Unknown Feedly Summary: QwQ: Reflect Deeply on the Boundaries of the Unknown Brand openly licensed model from Alibaba Cloud’s Qwen team, this time clearly inspired by OpenAI’s work on reasoning in o1. I love how the introduce the new…
-
Hacker News: Are Overemployed ‘Ghost Engineers’ Making Six Figures to Do Nothing?
Source URL: https://www.404media.co/are-overemployed-ghost-engineers-making-six-figures-to-do-nothing/ Source: Hacker News Title: Are Overemployed ‘Ghost Engineers’ Making Six Figures to Do Nothing? Feedly Summary: Comments AI Summary and Description: Yes Summary: The text discusses a viral tweet by Stanford researcher Yegor Denisov-Blanch regarding an algorithm that identifies “Ghost Engineers,” software engineers who perform minimally at tech companies, thus exposing a…
-
Hacker News: AMD Releases ROCm Version 6.3
Source URL: https://insidehpc.com/2024/11/amd-releases-rocm-version-6-3/ Source: Hacker News Title: AMD Releases ROCm Version 6.3 Feedly Summary: Comments AI Summary and Description: Yes Summary: AMD’s ROCm Version 6.3 enhances AI and HPC workloads through its advanced features like SGLang for generative AI, optimized FlashAttention-2, integration of the AMD Fortran compiler, and new multi-node FFT support. This release is…