Tag: model training
-
Simon Willison’s Weblog: Quoting Ben Thompson
Source URL: https://simonwillison.net/2025/Jan/28/ben-thompson/#atom-everything Source: Simon Willison’s Weblog Title: Quoting Ben Thompson Feedly Summary: H100s were prohibited by the chip ban, but not H800s. Everyone assumed that training leading edge models required more interchip memory bandwidth, but that is exactly what DeepSeek optimized both their model structure and infrastructure around. Again, just to emphasize this point,…
-
The Register: What happens when we can’t just build bigger AI datacenters anymore?
Source URL: https://www.theregister.com/2025/01/24/build_bigger_ai_datacenters/ Source: The Register Title: What happens when we can’t just build bigger AI datacenters anymore? Feedly Summary: We stitch together enormous supercomputers from other smaller supercomputers of course Feature Generative AI models have not only exploded in popularity over the past two years, but they’ve also grown at a precipitous rate, necessitating…
-
The Register: OpenAI’s Operator agent wants to tackle your online chores – just don’t expect it to nail every task
Source URL: https://www.theregister.com/2025/01/23/openai_unveils_operator_agent/ Source: The Register Title: OpenAI’s Operator agent wants to tackle your online chores – just don’t expect it to nail every task Feedly Summary: Hello Operator? Can you give me number nine? Can I see you later? Will you give me back my dime? OpenAI on Thursday launched a human-directed AI agent…
-
Hacker News: DeepSeek and the Effects of GPU Export Controls
Source URL: https://www.vincentschmalbach.com/deepseek-and-the-effects-of-gpu-export-controls/ Source: Hacker News Title: DeepSeek and the Effects of GPU Export Controls Feedly Summary: Comments AI Summary and Description: Yes Summary: DeepSeek’s unveiling of their V3 model demonstrates that AI advancements do not solely depend on high-end hardware but can be achieved through architectural efficiency. The model, trained on significantly fewer resources…