Tag: GKE
-
Cloud Blog: 65,000 nodes and counting: Google Kubernetes Engine is ready for trillion-parameter AI models
Source URL: https://cloud.google.com/blog/products/containers-kubernetes/gke-65k-nodes-and-counting/ Source: Cloud Blog Title: 65,000 nodes and counting: Google Kubernetes Engine is ready for trillion-parameter AI models Feedly Summary: As generative AI evolves, we’re beginning to see the transformative potential it is having across industries and our lives. And as large language models (LLMs) increase in size — current models are reaching…
-
Cloud Blog: Etsy’s Service Platform on Cloud Run cuts deployment time from days to under an hour
Source URL: https://cloud.google.com/blog/products/infrastructure/etsys-service-platform-on-cloud-run-cuts-deployment-time-from-days-to-minutes/ Source: Cloud Blog Title: Etsy’s Service Platform on Cloud Run cuts deployment time from days to under an hour Feedly Summary: Introduction Etsy, a leading ecommerce marketplace for handmade, vintage, and unique items has a passion for delivering innovative and seamless experiences for customers. Like many fast growing companies, Etsy needed to…
-
Cloud Blog: Powerful infrastructure innovations for your AI-first future
Source URL: https://cloud.google.com/blog/products/compute/trillium-sixth-generation-tpu-is-in-preview/ Source: Cloud Blog Title: Powerful infrastructure innovations for your AI-first future Feedly Summary: The rise of generative AI has ushered in an era of unprecedented innovation, demanding increasingly complex and more powerful AI models. These advanced models necessitate high-performance infrastructure capable of efficiently scaling AI training, tuning, and inferencing workloads while optimizing…
-
Cloud Blog: Save on GPUs: Smarter autoscaling for your GKE inferencing workloads
Source URL: https://cloud.google.com/blog/products/containers-kubernetes/tuning-the-gke-hpa-to-run-inference-on-gpus/ Source: Cloud Blog Title: Save on GPUs: Smarter autoscaling for your GKE inferencing workloads Feedly Summary: While LLM models deliver immense value for an increasing number of use cases, running LLM inference workloads can be costly. If you’re taking advantage of the latest open models and infrastructure, autoscaling can help you optimize…