Tag: plugin
- 
		
		
		
Cloud Blog: Now run your custom code at the edge with the Application Load Balancers
Source URL: https://cloud.google.com/blog/products/networking/service-extensions-plugins-for-application-load-balancers/ Source: Cloud Blog Title: Now run your custom code at the edge with the Application Load Balancers Feedly Summary: Application Load Balancers are essential for reliable web application delivery on Google Cloud. But while Google Cloud’s load balancers offer extensive customization, some situations demand even greater programmability. We recently announced Service Extensions…
 - 
		
		
		
Simon Willison’s Weblog: Claude 3.5 Haiku
Source URL: https://simonwillison.net/2024/Nov/4/haiku/#atom-everything Source: Simon Willison’s Weblog Title: Claude 3.5 Haiku Feedly Summary: Anthropic released Claude 3.5 Haiku today, a few days later than expected (they said it would be out by the end of October). I was expecting this to be a complete replacement for their existing Claude 3 Haiku model, in the same…
 - 
		
		
		
Hacker News: XTP: Make Squishy Software
Source URL: https://www.getxtp.com/blog/meet-xtp Source: Hacker News Title: XTP: Make Squishy Software Feedly Summary: Comments AI Summary and Description: Yes Summary: The XTP platform allows end-users to build and run plugins in a secure environment, enhancing the extensibility of applications. It utilizes WebAssembly (Wasm) for sandboxing, ensuring security even when executing potentially untrusted code. This innovation…
 - 
		
		
		
Hacker News: SmolLM2
Source URL: https://simonwillison.net/2024/Nov/2/smollm2/ Source: Hacker News Title: SmolLM2 Feedly Summary: Comments AI Summary and Description: Yes Summary: The text introduces SmolLM2, a new family of compact language models from Hugging Face, designed for lightweight on-device operations. The models, which range from 135M to 1.7B parameters, were trained on 11 trillion tokens across diverse datasets, showcasing…
 - 
		
		
		
Simon Willison’s Weblog: SmolLM2
Source URL: https://simonwillison.net/2024/Nov/2/smollm2/#atom-everything Source: Simon Willison’s Weblog Title: SmolLM2 Feedly Summary: SmolLM2 New from Loubna Ben Allal and her research team at Hugging Face: SmolLM2 is a family of compact language models available in three size: 135M, 360M, and 1.7B parameters. They are capable of solving a wide range of tasks while being lightweight enough…