Artwork

Contenu fourni par tinyML Foundation and TinyML Foundation. Tout le contenu du podcast, y compris les épisodes, les graphiques et les descriptions de podcast, est téléchargé et fourni directement par tinyML Foundation and TinyML Foundation ou son partenaire de plateforme de podcast. Si vous pensez que quelqu'un utilise votre œuvre protégée sans votre autorisation, vous pouvez suivre le processus décrit ici https://fr.player.fm/legal.
Player FM - Application Podcast
Mettez-vous hors ligne avec l'application Player FM !

Optimization Techniques for Powerful yet Tiny Machine Learning Models

59:37
 
Partager
 

Manage episode 422964912 series 3574631
Contenu fourni par tinyML Foundation and TinyML Foundation. Tout le contenu du podcast, y compris les épisodes, les graphiques et les descriptions de podcast, est téléchargé et fourni directement par tinyML Foundation and TinyML Foundation ou son partenaire de plateforme de podcast. Si vous pensez que quelqu'un utilise votre œuvre protégée sans votre autorisation, vous pouvez suivre le processus décrit ici https://fr.player.fm/legal.

Can machine learning models be both powerful and tiny? Join us in this episode of TinyML Talks, where we uncover groundbreaking techniques for making machine learning more efficient through high-level synthesis. We sit down with Russell Clayne, Technical Director at Siemens EDA, who guides us through the intricate process of pruning convolutional and deep neural networks. Discover how post-training quantization and quantization-aware training can trim down models without sacrificing performance, making them perfect for custom hardware accelerators like FPGAs and ASICs.
From there, we dive into a practical case study involving an MNIST-based network. Russell demonstrates how sensitivity analysis, network pruning, and quantization can significantly reduce neural network size while maintaining accuracy. Learn why fixed-point arithmetic is superior to floating-point in custom hardware, and how leading research from MIT and industry advancements are revolutionizing automated network optimization and model compression. You'll gain insights into how these techniques are not just theoretical but are being applied in real-world scenarios to save area and energy consumption.
Finally, explore the collaborative efforts between Siemens, Columbia University, and Global Foundries in a wake word analysis project. Russell explains how transitioning to hardware accelerators via high-level synthesis (HLS) tools can yield substantial performance improvements and energy savings. Understand the practicalities of using algorithmic C data types and Python-to-RTL tools to optimize ML workflows. Whether it's quantization-aware training, data movement optimization, or the fine details of using HLS libraries, this episode is packed with actionable insights for streamlining your machine learning models.

Learn more about the tinyML Foundation - tinyml.org

  continue reading

Chapitres

1. TinyML Talks (00:00:00)

2. Network Pruning and Quantization (00:10:51)

3. Optimizing Quantized Neural Networks (00:21:51)

4. High-Level Synthesis for ML Acceleration (00:37:27)

5. Hardware Design and Optimization Techniques (00:47:06)

5 episodes

Artwork
iconPartager
 
Manage episode 422964912 series 3574631
Contenu fourni par tinyML Foundation and TinyML Foundation. Tout le contenu du podcast, y compris les épisodes, les graphiques et les descriptions de podcast, est téléchargé et fourni directement par tinyML Foundation and TinyML Foundation ou son partenaire de plateforme de podcast. Si vous pensez que quelqu'un utilise votre œuvre protégée sans votre autorisation, vous pouvez suivre le processus décrit ici https://fr.player.fm/legal.

Can machine learning models be both powerful and tiny? Join us in this episode of TinyML Talks, where we uncover groundbreaking techniques for making machine learning more efficient through high-level synthesis. We sit down with Russell Clayne, Technical Director at Siemens EDA, who guides us through the intricate process of pruning convolutional and deep neural networks. Discover how post-training quantization and quantization-aware training can trim down models without sacrificing performance, making them perfect for custom hardware accelerators like FPGAs and ASICs.
From there, we dive into a practical case study involving an MNIST-based network. Russell demonstrates how sensitivity analysis, network pruning, and quantization can significantly reduce neural network size while maintaining accuracy. Learn why fixed-point arithmetic is superior to floating-point in custom hardware, and how leading research from MIT and industry advancements are revolutionizing automated network optimization and model compression. You'll gain insights into how these techniques are not just theoretical but are being applied in real-world scenarios to save area and energy consumption.
Finally, explore the collaborative efforts between Siemens, Columbia University, and Global Foundries in a wake word analysis project. Russell explains how transitioning to hardware accelerators via high-level synthesis (HLS) tools can yield substantial performance improvements and energy savings. Understand the practicalities of using algorithmic C data types and Python-to-RTL tools to optimize ML workflows. Whether it's quantization-aware training, data movement optimization, or the fine details of using HLS libraries, this episode is packed with actionable insights for streamlining your machine learning models.

Learn more about the tinyML Foundation - tinyml.org

  continue reading

Chapitres

1. TinyML Talks (00:00:00)

2. Network Pruning and Quantization (00:10:51)

3. Optimizing Quantized Neural Networks (00:21:51)

4. High-Level Synthesis for ML Acceleration (00:37:27)

5. Hardware Design and Optimization Techniques (00:47:06)

5 episodes

所有剧集

×
 
Loading …

Bienvenue sur Lecteur FM!

Lecteur FM recherche sur Internet des podcasts de haute qualité que vous pourrez apprécier dès maintenant. C'est la meilleure application de podcast et fonctionne sur Android, iPhone et le Web. Inscrivez-vous pour synchroniser les abonnements sur tous les appareils.

 

Guide de référence rapide