Tag: model inference
-
Simon Willison’s Weblog: llm-cerebras
Source URL: https://simonwillison.net/2024/Oct/25/llm-cerebras/ Source: Simon Willison’s Weblog Title: llm-cerebras Feedly Summary: llm-cerebras Cerebras (previously) provides Llama LLMs hosted on custom hardware at ferociously high speeds. GitHub user irthomasthomas built an LLM plugin that works against their API – which is currently free, albeit with a rate limit of 30 requests per minute for their two…
-
Hacker News: 1-Click Models Powered by Hugging Face
Source URL: https://www.digitalocean.com/blog/one-click-models-on-do-powered-by-huggingface Source: Hacker News Title: 1-Click Models Powered by Hugging Face Feedly Summary: Comments AI Summary and Description: Yes Summary: DigitalOcean has launched a new 1-Click Model deployment service powered by Hugging Face, termed HUGS on DO. This feature allows users to quickly deploy popular generative AI models on DigitalOcean GPU Droplets, aiming…
-
Simon Willison’s Weblog: lm.rs: run inference on Language Models locally on the CPU with Rust
Source URL: https://simonwillison.net/2024/Oct/11/lmrs/ Source: Simon Willison’s Weblog Title: lm.rs: run inference on Language Models locally on the CPU with Rust Feedly Summary: lm.rs: run inference on Language Models locally on the CPU with Rust Impressive new LLM inference implementation in Rust by Samuel Vitorino. I tried it just now on an M2 Mac with 64GB…