Tag: Hugging Face

  • Simon Willison’s Weblog: SmolVLM – small yet mighty Vision Language Model

    Source URL: https://simonwillison.net/2024/Nov/28/smolvlm/#atom-everything Source: Simon Willison’s Weblog Title: SmolVLM – small yet mighty Vision Language Model Feedly Summary: SmolVLM – small yet mighty Vision Language Model I’ve been having fun playing with this new vision model from the Hugging Face team behind SmolLM. They describe it as: […] a 2B VLM, SOTA for its memory…

  • Hacker News: Bluesky’s open API means anyone can scrape your data for AI training

    Source URL: https://techcrunch.com/2024/11/27/blueskys-open-api-means-anyone-can-scrape-your-data-for-ai-training/ Source: Hacker News Title: Bluesky’s open API means anyone can scrape your data for AI training Feedly Summary: Comments AI Summary and Description: Yes Summary: The text addresses concerns about user content privacy on the Bluesky social networking platform, particularly in relation to third-party access and machine learning data usage. It highlights…

  • Slashdot: OpenAI’s Sora Video Generator Appears To Have Leaked

    Source URL: https://slashdot.org/story/24/11/26/2020220/openais-sora-video-generator-appears-to-have-leaked?utm_source=rss1.0mainlinkanon&utm_medium=feed Source: Slashdot Title: OpenAI’s Sora Video Generator Appears To Have Leaked Feedly Summary: AI Summary and Description: Yes Summary: The text discusses a group that has leaked access to OpenAI’s Sora video generator, citing protests against perceived duplicity in OpenAI’s practices. This incident raises ongoing concerns about security in AI services and…

  • Hacker News: Open source audio transcription model that obscures sensitive info in realtime

    Source URL: https://venturebeat.com/ai/aiola-unveils-open-source-ai-audio-transcription-model-that-obscures-sensitive-info-in-realtime/ Source: Hacker News Title: Open source audio transcription model that obscures sensitive info in realtime Feedly Summary: Comments AI Summary and Description: Yes Summary: The text details a new open-source AI model, Whisper-NER from Israeli startup aiOla, designed for audio transcription while enhancing privacy through named entity recognition (NER). This model integrates…

  • Simon Willison’s Weblog: Pixtral Large

    Source URL: https://simonwillison.net/2024/Nov/18/pixtral-large/ Source: Simon Willison’s Weblog Title: Pixtral Large Feedly Summary: Pixtral Large New today from Mistral: Today we announce Pixtral Large, a 124B open-weights multimodal model built on top of Mistral Large 2. Pixtral Large is the second model in our multimodal family and demonstrates frontier-level image understanding. The weights are out on…

  • Cloud Blog: How to deploy Llama 3.2-1B-Instruct model with Google Cloud Run GPU

    Source URL: https://cloud.google.com/blog/products/ai-machine-learning/how-to-deploy-llama-3-2-1b-instruct-model-with-google-cloud-run/ Source: Cloud Blog Title: How to deploy Llama 3.2-1B-Instruct model with Google Cloud Run GPU Feedly Summary: As open-source large language models (LLMs) become increasingly popular, developers are looking for better ways to access new models and deploy them on Cloud Run GPU. That’s why Cloud Run now offers fully managed NVIDIA…

  • Cloud Blog: How to deploy and serve multi-host gen AI large open models over GKE

    Source URL: https://cloud.google.com/blog/products/ai-machine-learning/deploy-and-serve-open-models-over-google-kubernetes-engine/ Source: Cloud Blog Title: How to deploy and serve multi-host gen AI large open models over GKE Feedly Summary: Context As generative AI experiences explosive growth fueled by advancements in LLMs (Large Language Models), access to open models is more critical than ever for developers. Open models are publicly available pre-trained foundational…

  • Hacker News: Tencent drops a 389B MoE model(Open-source and free for commercial use))

    Source URL: https://github.com/Tencent/Tencent-Hunyuan-Large Source: Hacker News Title: Tencent drops a 389B MoE model(Open-source and free for commercial use)) Feedly Summary: Comments AI Summary and Description: Yes **Summary:** The text introduces the Hunyuan-Large model, the largest open-source Transformer-based Mixture of Experts (MoE) model, developed by Tencent, which boasts 389 billion parameters, optimizing performance while managing resource…

  • Simon Willison’s Weblog: Nous Hermes 3

    Source URL: https://simonwillison.net/2024/Nov/4/nous-hermes-3/#atom-everything Source: Simon Willison’s Weblog Title: Nous Hermes 3 Feedly Summary: Nous Hermes 3 The Nous Hermes family of fine-tuned models have a solid reputation. Their most recent release came out in August, based on Meta’s Llama 3.1: Our training data aggressively encourages the model to follow the system and instruction prompts exactly…

  • Hacker News: SmolLM2

    Source URL: https://simonwillison.net/2024/Nov/2/smollm2/ Source: Hacker News Title: SmolLM2 Feedly Summary: Comments AI Summary and Description: Yes Summary: The text introduces SmolLM2, a new family of compact language models from Hugging Face, designed for lightweight on-device operations. The models, which range from 135M to 1.7B parameters, were trained on 11 trillion tokens across diverse datasets, showcasing…