Tag: parameter

  • The Register: DeepSeek-R1-beating perf in a 32B package? El Reg digs its claws into Alibaba’s QwQ

    Source URL: https://www.theregister.com/2025/03/16/qwq_hands_on_review/ Source: The Register Title: DeepSeek-R1-beating perf in a 32B package? El Reg digs its claws into Alibaba’s QwQ Feedly Summary: How to tame its hypersensitive hyperparameters and get it running on your PC Hands on How much can reinforcement learning – and a bit of extra verification – improve large language models,…

  • Simon Willison’s Weblog: mlx-community/OLMo-2-0325-32B-Instruct-4bit

    Source URL: https://simonwillison.net/2025/Mar/16/olmo2/#atom-everything Source: Simon Willison’s Weblog Title: mlx-community/OLMo-2-0325-32B-Instruct-4bit Feedly Summary: mlx-community/OLMo-2-0325-32B-Instruct-4bit OLMo 2 32B claims to be “the first fully-open model (all data, code, weights, and details are freely available) to outperform GPT3.5-Turbo and GPT-4o mini". Thanks to the MLX project here’s a recipe that worked for me to run it on my Mac,…

  • Hacker News: AutoHete: An Automatic and Efficient Heterogeneous Training System for LLMs

    Source URL: https://arxiv.org/abs/2503.01890 Source: Hacker News Title: AutoHete: An Automatic and Efficient Heterogeneous Training System for LLMs Feedly Summary: Comments AI Summary and Description: Yes Summary: The paper introduces AutoHete, a groundbreaking training system designed for heterogeneous environments that significantly enhances the training efficiency of large language models (LLMs). It addresses GPU memory limitations and…

  • Simon Willison’s Weblog: My tools colophon now has AI-generated descriptions

    Source URL: https://simonwillison.net/2025/Mar/13/tools-colophon/ Source: Simon Willison’s Weblog Title: My tools colophon now has AI-generated descriptions Feedly Summary: My tools colophon now has AI-generated descriptions The /colophon page on my tools site lists all 78 of my tools along with their commit histories, including links to prompting transcripts. I wrote about how I built that. the…

  • Hacker News: Gemma3 – The current strongest model that fits on a single GPU

    Source URL: https://ollama.com/library/gemma3 Source: Hacker News Title: Gemma3 – The current strongest model that fits on a single GPU Feedly Summary: Comments AI Summary and Description: Yes **Summary:** The text discusses the features and capabilities of the Gemma 3 models developed by Google, which are built on Gemini technology and designed for multimodal tasks. Their…

  • Hacker News: Gemma 3 Technical Report [pdf]

    Source URL: https://storage.googleapis.com/deepmind-media/gemma/Gemma3Report.pdf Source: Hacker News Title: Gemma 3 Technical Report [pdf] Feedly Summary: Comments AI Summary and Description: Yes **Summary:** The text provides a comprehensive technical report on Gemma 3, an advanced multimodal language model introduced by Google DeepMind. It highlights significant architectural improvements, including an increased context size, enhanced multilingual capabilities, and innovations…

  • Hacker News: A Practical Guide to Running Local LLMs

    Source URL: https://spin.atomicobject.com/running-local-llms/ Source: Hacker News Title: A Practical Guide to Running Local LLMs Feedly Summary: Comments AI Summary and Description: Yes Summary: The text discusses the intricacies of running local large language models (LLMs), emphasizing their applications in privacy-critical situations and the potential benefits of various tools like Ollama and Llama.cpp. It provides insights…

  • Hacker News: Smaller but Better: Unifying Layout Generation with Smaller LLMs

    Source URL: https://arxiv.org/abs/2502.14005 Source: Hacker News Title: Smaller but Better: Unifying Layout Generation with Smaller LLMs Feedly Summary: Comments AI Summary and Description: Yes Summary: The paper presents LGGPT, a large language model designed for unified layout generation, emphasizing its efficiency and performance even with a smaller size compared to larger models. It introduces novel…