Hardware acceleration for on-device Machine Learning

6 592
21.5
Следующее
Популярные
76 дней – 3 8530:45
😎 Shaders in Jetpack Compose
Опубликовано 15 ноября 2022, 17:05
Hardware acceleration can dramatically reduce inference latency for machine learning enabled features and allow you to deliver live on-device experiences that may not be possible otherwise.
Today, in addition to CPU, Android devices embed various specialized chips such as GPU, DSP or NPU that you can use to accelerate your ML inference.
In this talk we go over some tools and solutions offered by TensorFlow and Android ML teams that help you take advantage of various hardware to accelerate ML inference in your Android app.

Resources:
TensorFlow documentation→ goo.gle/3UCuw2L
GPU delegate documentation → goo.gle/3DQMWGe
Model analyzer → goo.gle/3NRuKAN
NNAPI delegate documentation: → goo.gle/3tc4ibB
Performance delegates documentation → goo.gle/3TiZeNd
Acceleration Service → goo.gle/3hkxMRT
Android ML documentation → goo.gle/3tbzcko

Speaker: Thomas Ezan

Watch more:
Watch all the Android Dev Summit sessions → goo.gle/ADS-All
Watch all the Platform track sessions → goo.gle/ADS-Platform

Subscribe to Android Developers → goo.gle/AndroidDevs

#Featured #AndroidDevSummit #Android
автотехномузыкадетское