Tag: transformers-js

  • Simon Willison’s Weblog: Load Llama-3.2 WebGPU in your browser from a local folder

    Source URL: https://simonwillison.net/2025/Sep/8/webgpu-local-folder/#atom-everything Source: Simon Willison’s Weblog Title: Load Llama-3.2 WebGPU in your browser from a local folder Feedly Summary: Load Llama-3.2 WebGPU in your browser from a local folder Inspired by a comment on Hacker News I decided to see if it was possible to modify the transformers.js-examples/tree/main/llama-3.2-webgpu Llama 3.2 chat demo (online here,…

  • Simon Willison’s Weblog: Introducing EmbeddingGemma

    Source URL: https://simonwillison.net/2025/Sep/4/embedding-gemma/#atom-everything Source: Simon Willison’s Weblog Title: Introducing EmbeddingGemma Feedly Summary: Introducing EmbeddingGemma Brand new open weights (under the slightly janky Gemma license) 308M parameter embedding model from Google: Based on the Gemma 3 architecture, EmbeddingGemma is trained on 100+ languages and is small enough to run on less than 200MB of RAM with…

  • Simon Willison’s Weblog: llama-3.2-webgpu

    Source URL: https://simonwillison.net/2024/Sep/30/llama-32-webgpu/#atom-everything Source: Simon Willison’s Weblog Title: llama-3.2-webgpu Feedly Summary: llama-3.2-webgpu Llama 3.2 1B is a really interesting models, given its 128,000 token input and its tiny size (barely more than a GB). This page loads a 1.24GB q4f16 ONNX build of the Llama-3.2-1B-Instruct model and runs it with a React-powered chat interface directly…