Hardware acceleration for on-device Machine Learning

6 076
20.7
Опубликовано 15 ноября 2022, 17:05
Hardware acceleration can dramatically reduce inference latency for machine learning enabled features and allow you to deliver live on-device experiences that may not be possible otherwise.
Today, in addition to CPU, Android devices embed various specialized chips such as GPU, DSP or NPU that you can use to accelerate your ML inference.
In this talk we go over some tools and solutions offered by TensorFlow and Android ML teams that help you take advantage of various hardware to accelerate ML inference in your Android app.

Resources:
TensorFlow documentation→ goo.gle/3UCuw2L
GPU delegate documentation → goo.gle/3DQMWGe
Model analyzer → goo.gle/3NRuKAN
NNAPI delegate documentation: → goo.gle/3tc4ibB
Performance delegates documentation → goo.gle/3TiZeNd
Acceleration Service → goo.gle/3hkxMRT
Android ML documentation → goo.gle/3tbzcko

Speaker: Thomas Ezan

Watch more:
Watch all the Android Dev Summit sessions → goo.gle/ADS-All
Watch all the Platform track sessions → goo.gle/ADS-Platform

Subscribe to Android Developers → goo.gle/AndroidDevs

#Featured #AndroidDevSummit #Android
Случайные видео
187 дней – 65 6400:57
OnePlus Open Unboxing
321 день – 1 731 7495:11
Bespoke Life 2023 Highlights l Samsung
07.10.21 – 5 865 1842:57
Celebrating Steve | October 5 | Apple
автотехномузыкадетское