Tag: Hugging Face

  • Hacker News: Tencent drops a 389B MoE model(Open-source and free for commercial use))

    Source URL: https://github.com/Tencent/Tencent-Hunyuan-Large Source: Hacker News Title: Tencent drops a 389B MoE model(Open-source and free for commercial use)) Feedly Summary: Comments AI Summary and Description: Yes **Summary:** The text introduces the Hunyuan-Large model, the largest open-source Transformer-based Mixture of Experts (MoE) model, developed by Tencent, which boasts 389 billion parameters, optimizing performance while managing resource…

  • Simon Willison’s Weblog: Nous Hermes 3

    Source URL: https://simonwillison.net/2024/Nov/4/nous-hermes-3/#atom-everything Source: Simon Willison’s Weblog Title: Nous Hermes 3 Feedly Summary: Nous Hermes 3 The Nous Hermes family of fine-tuned models have a solid reputation. Their most recent release came out in August, based on Meta’s Llama 3.1: Our training data aggressively encourages the model to follow the system and instruction prompts exactly…

  • Hacker News: SmolLM2

    Source URL: https://simonwillison.net/2024/Nov/2/smollm2/ Source: Hacker News Title: SmolLM2 Feedly Summary: Comments AI Summary and Description: Yes Summary: The text introduces SmolLM2, a new family of compact language models from Hugging Face, designed for lightweight on-device operations. The models, which range from 135M to 1.7B parameters, were trained on 11 trillion tokens across diverse datasets, showcasing…

  • Simon Willison’s Weblog: SmolLM2

    Source URL: https://simonwillison.net/2024/Nov/2/smollm2/#atom-everything Source: Simon Willison’s Weblog Title: SmolLM2 Feedly Summary: SmolLM2 New from Loubna Ben Allal and her research team at Hugging Face: SmolLM2 is a family of compact language models available in three size: 135M, 360M, and 1.7B parameters. They are capable of solving a wide range of tasks while being lightweight enough…

  • Cloud Blog: PyTorch/XLA 2.5: vLLM support and an improved developer experience

    Source URL: https://cloud.google.com/blog/products/ai-machine-learning/whats-new-with-pytorchxla-2-5/ Source: Cloud Blog Title: PyTorch/XLA 2.5: vLLM support and an improved developer experience Feedly Summary: Machine learning engineers are bullish on PyTorch/XLA, a Python package that uses the XLA deep learning compiler to connect the PyTorch deep learning framework and Cloud TPUs. And now, PyTorch/XLA 2.5 is here, along with a set…

  • Hacker News: Meta’s Open Source NotebookLM

    Source URL: https://github.com/meta-llama/llama-recipes/tree/main/recipes/quickstart/NotebookLlama Source: Hacker News Title: Meta’s Open Source NotebookLM Feedly Summary: Comments AI Summary and Description: Yes Summary: The text presents a comprehensive guide to using an open-source project called NotebookLlama, aimed at creating a workflow that converts PDF documents into podcasts using various LLMs (Large Language Models). This process is likely to…

  • The Register: Hugging Face puts the squeeze on Nvidia’s software ambitions

    Source URL: https://www.theregister.com/2024/10/24/huggingface_hugs_nvidia/ Source: The Register Title: Hugging Face puts the squeeze on Nvidia’s software ambitions Feedly Summary: AI model repo promises lower costs, broader compatibility for NIMs competitor Hugging Face this week announced HUGS, its answer to Nvidia’s Inference Microservices (NIMs), which the AI repo claims will let customers deploy and run LLMs and…

  • Hacker News: 1-Click Models Powered by Hugging Face

    Source URL: https://www.digitalocean.com/blog/one-click-models-on-do-powered-by-huggingface Source: Hacker News Title: 1-Click Models Powered by Hugging Face Feedly Summary: Comments AI Summary and Description: Yes Summary: DigitalOcean has launched a new 1-Click Model deployment service powered by Hugging Face, termed HUGS on DO. This feature allows users to quickly deploy popular generative AI models on DigitalOcean GPU Droplets, aiming…

  • Hacker News: .txt raises $11.9M to make language models programmable

    Source URL: https://techcrunch.com/2024/10/17/with-11-9-million-in-funding-dottxt-tells-ai-models-how-to-answer/ Source: Hacker News Title: .txt raises $11.9M to make language models programmable Feedly Summary: Comments AI Summary and Description: Yes Summary: The text discusses Dottxt, a U.S.-based startup that aims to improve the interaction between generative AI models and traditional software engineering workflows. By utilizing structured generation methods, Dottxt helps developers obtain…

  • Hacker News: Microsoft BitNet: inference framework for 1-bit LLMs

    Source URL: https://github.com/microsoft/BitNet Source: Hacker News Title: Microsoft BitNet: inference framework for 1-bit LLMs Feedly Summary: Comments AI Summary and Description: Yes Summary: The text describes “bitnet.cpp,” a specialized inference framework for 1-bit large language models (LLMs), specifically highlighting its performance enhancements, optimized kernel support, and installation instructions. This framework is poised to significantly influence…