Tag: language

  • Hamel’s Blog: LLM Eval FAQ

    Source URL: https://hamel.dev/blog/posts/evals-faq/ Source: Hamel’s Blog Title: LLM Eval FAQ Feedly Summary: Our Course On AI Evals I’m teaching a course on AI Evals with Shreya Shankar. Here are some of the most common questions we’ve been asked. We’ll be updating this list frequently. Q: Is RAG dead? Question: Should I avoid using RAG for…

  • Simon Willison’s Weblog: llm-tools-exa

    Source URL: https://simonwillison.net/2025/May/29/llm-tools-exa/ Source: Simon Willison’s Weblog Title: llm-tools-exa Feedly Summary: llm-tools-exa When I shipped LLM 0.26 yesterday one of the things I was most excited about was seeing what new tool plugins people would build for it. Dan Turkel’s llm-tools-exa is one of the first. It adds web search to LLM using Exa (previously),…

  • Simon Willison’s Weblog: AI-assisted development needs automated tests

    Source URL: https://simonwillison.net/2025/May/28/automated-tests/ Source: Simon Willison’s Weblog Title: AI-assisted development needs automated tests Feedly Summary: I wonder if one of the reasons I’m finding LLMs so much more useful for coding than a lot of people that I see in online discussions is that effectively all of the code I work on has automated tests.…

  • Simon Willison’s Weblog: llm-llama-server 0.2

    Source URL: https://simonwillison.net/2025/May/28/llama-server-tools/ Source: Simon Willison’s Weblog Title: llm-llama-server 0.2 Feedly Summary: llm-llama-server 0.2 Here’s a second option for using LLM’s new tool support against local models (the first was via llm-ollama). It turns out the llama.cpp ecosystem has pretty robust OpenAI-compatible tool support already, so my llm-llama-server plugin only needed a quick upgrade to…

  • Simon Willison’s Weblog: Large Language Models can run tools in your terminal with LLM 0.26

    Source URL: https://simonwillison.net/2025/May/27/llm-tools/ Source: Simon Willison’s Weblog Title: Large Language Models can run tools in your terminal with LLM 0.26 Feedly Summary: LLM 0.26 is out with the biggest new feature since I started the project: support for tools. You can now use the LLM CLI tool – and Python library – to grant LLMs…

  • AWS News Blog: Amazon Aurora DSQL is now generally available

    Source URL: https://aws.amazon.com/blogs/aws/amazon-aurora-dsql-is-now-generally-available/ Source: AWS News Blog Title: Amazon Aurora DSQL is now generally available Feedly Summary: Amazon Aurora DSQL is the fastest serverless distributed SQL database for always available applications. It makes it effortless for customers to scale to meet any workload demand with zero infrastructure management and zero downtime maintenance. With its active-active…

  • Simon Willison’s Weblog: Build AI agents with the Mistral Agents API

    Source URL: https://simonwillison.net/2025/May/27/mistral-agents-api/ Source: Simon Willison’s Weblog Title: Build AI agents with the Mistral Agents API Feedly Summary: Build AI agents with the Mistral Agents API Big upgrade to Mistral’s API this morning: they’ve announced a new “Agents API". Mistral have been using the term "agents" for a while now. Here’s how they describe them:…

  • Scott Logic: Read the books! Should junior developers use LLMs?

    Source URL: https://blog.scottlogic.com/2025/05/27/read-the-books-should-junior-developers-use-llms.html Source: Scott Logic Title: Read the books! Should junior developers use LLMs? Feedly Summary: Large Language Models are powerful tools that can greatly enhance software developers’ productivity, but for junior developers starting a career in tech, they may hinder long-term growth by abstracting away essential programming fundamentals. AI Summary and Description: Yes…

  • Simon Willison’s Weblog: GitHub MCP Exploited: Accessing private repositories via MCP

    Source URL: https://simonwillison.net/2025/May/26/github-mcp-exploited/#atom-everything Source: Simon Willison’s Weblog Title: GitHub MCP Exploited: Accessing private repositories via MCP Feedly Summary: GitHub MCP Exploited: Accessing private repositories via MCP GitHub’s official MCP server grants LLMs a whole host of new abilities, including being able to read and issues in repositories the user has access to and submit new…