Accelerating ML development with optimized performance and cost

979
17.2
Опубликовано 11 октября 2022, 17:01
Successful machine learning deployments can depend on the ability of infrastructure to support the performance and budget requirements of a workload. Google’s open, flexible, and scalable AI infrastructure supports a wide variety of AI workloads, enabling customers to increase velocity to production, reduce costs, and to meet changing requirements over time. In this session, we discuss how to optimize across the AI stack with the latest GPUs and TPUs, fully-managed purpose-built AI infrastructure capabilities with Vertex AI, and state of the art solutions for demanding workloads. Hear about how Uber, Cohere, Credit Karma, Arbor Biotechnologies, and other enterprises are leveraging Google Cloud's AI Infrastructure to innovate, accelerate deployment, and drive business value.

Optimize training performance with Reduction Server on Vertex AI → goo.gle/3SE2VNX

Speakers: Mikhail Chrestkha, Kai Wang, Joanna Yoo, Siddhartha Kamalakara

Watch more:
All sessions from Google Cloud Next → goo.gle/next22-allsessions
Watch all the Modernize sessions → goo.gle/next22-modernize

Subscribe to Google Cloud Tech → goo.gle/GoogleCloudTech

#GoogleCloudNext

MOD300
Свежие видео
10 дней – 26 2251:51
What is Gemma Scope?
13 дней – 3 0344:50
Semantic modeling for AI
18 дней – 3 7715:48
Google Trends for Researchers
автотехномузыкадетское