Tag: machine learning workloads
-
Cloud Blog: Taming the stragglers: Maximize AI training performance with automated straggler detection
Source URL: https://cloud.google.com/blog/products/compute/stragglers-in-ai-a-guide-to-automated-straggler-detection/ Source: Cloud Blog Title: Taming the stragglers: Maximize AI training performance with automated straggler detection Feedly Summary: Stragglers are an industry-wide issue for developers working with large-scale machine learning workloads. The larger and more powerful these systems become, the more their performance is hostage to the subtle misbehavior of a single component.…
-
Cloud Blog: Understanding Calendar mode for Dynamic Workload Scheduler: Reserve ML GPUs and TPUs
Source URL: https://cloud.google.com/blog/products/compute/dynamic-workload-scheduler-calendar-mode-reserves-gpus-and-tpus/ Source: Cloud Blog Title: Understanding Calendar mode for Dynamic Workload Scheduler: Reserve ML GPUs and TPUs Feedly Summary: Organizations need ML compute resources that can accommodate bursty peaks and periodic troughs. That means the consumption models for AI infrastructure need to evolve to be more cost-efficient, provide term flexibility, and support rapid…
-
The Register: Google offers 7th-gen Ironwood TPUs for AI, with AI-inspired comparisons
Source URL: https://www.theregister.com/2025/04/10/googles_7thgen_ironwood_tpus_debut/ Source: The Register Title: Google offers 7th-gen Ironwood TPUs for AI, with AI-inspired comparisons Feedly Summary: Sure, we’re doing FP8 versus a supercomputer’s FP64. What of it? Cloud Next Google’s seventh-generation Tensor Processing Units (TPU), announced Wednesday, will soon be available to cloud customers to rent in pods of 256 or 9,216…
-
Cloud Blog: Anyscale powers AI compute for any workload using Google Compute Engine
Source URL: https://cloud.google.com/blog/products/ai-machine-learning/anyscale-powers-ai-compute-for-any-workload-using-google-compute-engine/ Source: Cloud Blog Title: Anyscale powers AI compute for any workload using Google Compute Engine Feedly Summary: Over the past decade, AI has evolved at a breakneck pace, turning from a futuristic dream into a tool now accessible to everyone. One of the technologies that opened up this new era of AI…
-
Cloud Blog: Using RDMA over Converged Ethernet networking for AI on Google Cloud
Source URL: https://cloud.google.com/blog/products/networking/rdma-rocev2-for-ai-workloads-on-google-cloud/ Source: Cloud Blog Title: Using RDMA over Converged Ethernet networking for AI on Google Cloud Feedly Summary: All workloads are not the same. This is especially the case for AI, ML, and scientific workloads. In this blog we show how Google Cloud makes the RDMA over converged ethernet version 2 (RoCE v2)…