Tag: llms

  • New York Times – Artificial Intelligence : Can a Start-Up Help Authors Get Paid by A.I. Companies?

    Source URL: https://www.nytimes.com/2024/10/09/books/authors-guild-created-by-humans-artificial-intelligence.html Source: New York Times – Artificial Intelligence Title: Can a Start-Up Help Authors Get Paid by A.I. Companies? Feedly Summary: Created by Humans, a company that aims to help writers license their works for use by A.I. companies, has struck a partnership with the Authors Guild. AI Summary and Description: Yes Summary:…

  • Slashdot: Researchers Claim New Technique Slashes AI Energy Use By 95%

    Source URL: https://science.slashdot.org/story/24/10/08/2035247/researchers-claim-new-technique-slashes-ai-energy-use-by-95?utm_source=rss1.0mainlinkanon&utm_medium=feed Source: Slashdot Title: Researchers Claim New Technique Slashes AI Energy Use By 95% Feedly Summary: AI Summary and Description: Yes Summary: Researchers at BitEnergy AI, Inc. have introduced Linear-Complexity Multiplication (L-Mul), a novel technique that reduces AI model power consumption by up to 95% by replacing floating-point multiplications with integer additions. This…

  • Simon Willison’s Weblog: openai/openai-realtime-console

    Source URL: https://simonwillison.net/2024/Oct/9/openai-realtime-console/#atom-everything Source: Simon Willison’s Weblog Title: openai/openai-realtime-console Feedly Summary: openai/openai-realtime-console I got this OpenAI demo repository working today – it’s an extremely easy way to get started playing around with the new Realtime voice API they announced at DevDay last week: cd /tmp git clone https://github.com/openai/openai-realtime-console cd openai-realtime-console npm i npm start That…

  • Simon Willison’s Weblog: Anthropic: Message Batches (beta)

    Source URL: https://simonwillison.net/2024/Oct/8/anthropic-batch-mode/ Source: Simon Willison’s Weblog Title: Anthropic: Message Batches (beta) Feedly Summary: Anthropic: Message Batches (beta) Anthropic now have a batch mode, allowing you to send prompts to Claude in batches which will be processed within 24 hours (though probably much faster than that) and come at a 50% price discount. This matches…

  • Simon Willison’s Weblog: Gemini 1.5 Flash-8B is now production ready

    Source URL: https://simonwillison.net/2024/Oct/3/gemini-15-flash-8b/#atom-everything Source: Simon Willison’s Weblog Title: Gemini 1.5 Flash-8B is now production ready Feedly Summary: Gemini 1.5 Flash-8B is now production ready Gemini 1.5 Flash-8B is “a smaller and faster variant of 1.5 Flash" – and is now released to production, at half the price of the 1.5 Flash model. It’s really, really…

  • Simon Willison’s Weblog: OpenAI DevDay: Let’s build developer tools, not digital God

    Source URL: https://simonwillison.net/2024/Oct/2/not-digital-god/#atom-everything Source: Simon Willison’s Weblog Title: OpenAI DevDay: Let’s build developer tools, not digital God Feedly Summary: I had a fun time live blogging OpenAI DevDay yesterday – I’ve now shared notes about the live blogging system I threw other in a hurry on the day (with assistance from Claude and GPT-4o). Now…

  • Simon Willison’s Weblog: llama-3.2-webgpu

    Source URL: https://simonwillison.net/2024/Sep/30/llama-32-webgpu/#atom-everything Source: Simon Willison’s Weblog Title: llama-3.2-webgpu Feedly Summary: llama-3.2-webgpu Llama 3.2 1B is a really interesting models, given its 128,000 token input and its tiny size (barely more than a GB). This page loads a 1.24GB q4f16 ONNX build of the Llama-3.2-1B-Instruct model and runs it with a React-powered chat interface directly…

  • Simon Willison’s Weblog: Updated production-ready Gemini models

    Source URL: https://simonwillison.net/2024/Sep/24/gemini-models/#atom-everything Source: Simon Willison’s Weblog Title: Updated production-ready Gemini models Feedly Summary: Updated production-ready Gemini models Two new models from Google Gemini today: gemini-1.5-pro-002 and gemini-1.5-flash-002. Their -latest aliases will update to these new models in “the next few days", and new -001 suffixes can be used to stick with the older models.…

  • Simon Willison’s Weblog: How streaming LLM APIs work

    Source URL: https://simonwillison.net/2024/Sep/22/how-streaming-llm-apis-work/#atom-everything Source: Simon Willison’s Weblog Title: How streaming LLM APIs work Feedly Summary: How streaming LLM APIs work New TIL. I used curl to explore the streaming APIs provided by OpenAI, Anthropic and Google Gemini and wrote up detailed notes on what I learned. Tags: apis, http, llms, json AI Summary and Description:…