Google Developers2.51M
Published on 4 Jun 2025, 20:00
Discover how to build efficient edge AI applications with Google tools using Hybrid LLMs. Explore on-device and cloud-based models, leveraging both Gemma and Gemini through sequential, parallel and routing architectures. Optimize queries for performance, privacy, and model capabilities to meet your app needs..
Resources:
Discover Gemma capabilities→ goo.gle/try-gemma
Gemma Cookbook → goo.gle/gemma-cookbook
Google AI Studio → goo.gle/aistudio
Get started with the Gemini API → goo.gle/gemini-get-started
MediaPipe samples → goo.gle/gemma3-mediapipe-llm
Run Gemma with Ollama → goo.gle/gemma3-ollama
Serving Gemma 3 with Ollama on Cloud Run → goo.gle/gemma3-cloudrun-ollama
Serving Gemma 3 with vLLM on Cloud Run → goo.gle/gemma3-cloudrun-vllm
Subscribe to Google for Developers → goo.gle/developers
Speaker: Ian Ballantyne
Products mentioned: Gemma, Gemini
Resources:
Discover Gemma capabilities→ goo.gle/try-gemma
Gemma Cookbook → goo.gle/gemma-cookbook
Google AI Studio → goo.gle/aistudio
Get started with the Gemini API → goo.gle/gemini-get-started
MediaPipe samples → goo.gle/gemma3-mediapipe-llm
Run Gemma with Ollama → goo.gle/gemma3-ollama
Serving Gemma 3 with Ollama on Cloud Run → goo.gle/gemma3-cloudrun-ollama
Serving Gemma 3 with vLLM on Cloud Run → goo.gle/gemma3-cloudrun-vllm
Subscribe to Google for Developers → goo.gle/developers
Speaker: Ian Ballantyne
Products mentioned: Gemma, Gemini
Fresh videos
Random videos