Tag: DeepSeek

  • Simon Willison’s Weblog: Shisa V2 405B: Japan’s Highest Performing LLM

    Source URL: https://simonwillison.net/2025/Jun/3/shisa-v2/ Source: Simon Willison’s Weblog Title: Shisa V2 405B: Japan’s Highest Performing LLM Feedly Summary: Shisa V2 405B: Japan’s Highest Performing LLM Leonard Lin and Adam Lensenmayer have been working on Shisa for a while. They describe their latest release as “Japan’s Highest Performing LLM". Shisa V2 405B is the highest-performing LLM ever…

  • Slashdot: Web-Scraping AI Bots Cause Disruption For Scientific Databases and Journals

    Source URL: https://science.slashdot.org/story/25/06/02/172202/web-scraping-ai-bots-cause-disruption-for-scientific-databases-and-journals?utm_source=rss1.0mainlinkanon&utm_medium=feed Source: Slashdot Title: Web-Scraping AI Bots Cause Disruption For Scientific Databases and Journals Feedly Summary: AI Summary and Description: Yes Summary: The text highlights the impact of automated web-scraping bots on scientific databases and academic journals, driven by the demand for training data for AI models. This has led to significant service…

  • Simon Willison’s Weblog: How often do LLMs snitch? Recreating Theo’s SnitchBench with LLM

    Source URL: https://simonwillison.net/2025/May/31/snitchbench-with-llm/#atom-everything Source: Simon Willison’s Weblog Title: How often do LLMs snitch? Recreating Theo’s SnitchBench with LLM Feedly Summary: A fun new benchmark just dropped! Inspired by the Claude 4 system card – which showed that Claude 4 might just rat you out to the authorities if you told it to “take initiative" in…

  • Simon Willison’s Weblog: deepseek-ai/DeepSeek-R1-0528

    Source URL: https://simonwillison.net/2025/May/31/deepseek-aideepseek-r1-0528/ Source: Simon Willison’s Weblog Title: deepseek-ai/DeepSeek-R1-0528 Feedly Summary: deepseek-ai/DeepSeek-R1-0528 Sadly the trend for terrible naming of models has infested the Chinese AI labs as well. DeepSeek-R1-0528 is a brand new and much improved open weights reasoning model from DeepSeek, a major step up from the DeepSeek R1 they released back in January.…

  • Simon Willison’s Weblog: Talking AI and jobs with Natasha Zouves for News Nation

    Source URL: https://simonwillison.net/2025/May/30/ai-and-jobs-with-natasha-zouves/#atom-everything Source: Simon Willison’s Weblog Title: Talking AI and jobs with Natasha Zouves for News Nation Feedly Summary: I was interviewed by News Nation’s Natasha Zouves about the very complicated topic of how we should think about AI in terms of threatening our jobs and careers. I previously talked with Natasha two years…

  • Slashdot: Researchers Warn Against Treating AI Outputs as Human-Like Reasoning

    Source URL: https://tech.slashdot.org/story/25/05/29/1411236/researchers-warn-against-treating-ai-outputs-as-human-like-reasoning?utm_source=rss1.0mainlinkanon&utm_medium=feed Source: Slashdot Title: Researchers Warn Against Treating AI Outputs as Human-Like Reasoning Feedly Summary: AI Summary and Description: Yes Summary: Researchers at Arizona State University are challenging the misconception of AI language models’ intermediate outputs as “reasoning” or “thinking.” They argue that this anthropomorphization can mislead users about AI’s actual functioning, highlighting…

  • Simon Willison’s Weblog: Devstral

    Source URL: https://simonwillison.net/2025/May/21/devstral/#atom-everything Source: Simon Willison’s Weblog Title: Devstral Feedly Summary: Devstral New Apache 2.0 licensed LLM release from Mistral, this time specifically trained for code. Devstral achieves a score of 46.8% on SWE-Bench Verified, outperforming prior open-source SoTA models by more than 6% points. When evaluated under the same test scaffold (OpenHands, provided by…

  • Cloud Blog: AI Hypercomputer developer experience enhancements from Q1 25: build faster, scale bigger

    Source URL: https://cloud.google.com/blog/products/compute/ai-hypercomputer-enhancements-for-the-developer/ Source: Cloud Blog Title: AI Hypercomputer developer experience enhancements from Q1 25: build faster, scale bigger Feedly Summary: Building cutting-edge AI models is exciting, whether you’re iterating in your notebook or orchestrating large clusters. However, scaling up training can present significant challenges, including navigating complex infrastructure, configuring software and dependencies across numerous…

  • Simon Willison’s Weblog: Gemini 2.5 Models now support implicit caching

    Source URL: https://simonwillison.net/2025/May/9/gemini-implicit-caching/#atom-everything Source: Simon Willison’s Weblog Title: Gemini 2.5 Models now support implicit caching Feedly Summary: Gemini 2.5 Models now support implicit caching I just spotted a cacheTokensDetails key in the token usage JSON while running a long chain of prompts against Gemini 2.5 Flash – despite not configuring caching myself: {“cachedContentTokenCount": 200658, "promptTokensDetails":…

  • Simon Willison’s Weblog: What people get wrong about the leading Chinese open models: Adoption and censorship

    Source URL: https://simonwillison.net/2025/May/6/what-people-get-wrong-about-the-leading-chinese-models/#atom-everything Source: Simon Willison’s Weblog Title: What people get wrong about the leading Chinese open models: Adoption and censorship Feedly Summary: What people get wrong about the leading Chinese open models: Adoption and censorship While I’ve been enjoying trying out Alibaba’s Qwen 3 a lot recently, Nathan Lambert focuses on the elephant in…