Difference between revisions of "ML-TN-001 - AI at the edge: comparison of different embedded platforms - Part 1"

From DAVE Developer's Wiki
Jump to: navigation, search
(Introduction)
Line 19: Line 19:
  
 
The idea is to develop one or more applications with well-known open-source frameworks/libraries and to deploy them on such platforms to compare performances, resource utilization, development flow, etc.
 
The idea is to develop one or more applications with well-known open-source frameworks/libraries and to deploy them on such platforms to compare performances, resource utilization, development flow, etc.
 +
 +
==Test application #1: fruit classifier==
 +
This application implements a classifier like the one described [[MISC-TN-011:_Running_an_Azure-generated_TensorFlow_Lite_model_on_Mito8M_SoM_using_NXP_eIQ here]]. There is one notable difference, however. The model
 +
 +
TBD

Revision as of 13:55, 4 September 2020

Info Box
NeuralNetwork.png Applies to Machine Learning


History[edit | edit source]

Version Date Notes
1.0.0 September 2020 First public release

Introduction[edit | edit source]

This Technical Note (TN for short) is the first one of a series illustrating how machine learning-based inference applications perform across different embedded platforms, which are eligible for building intelligent edge devices.

The idea is to develop one or more applications with well-known open-source frameworks/libraries and to deploy them on such platforms to compare performances, resource utilization, development flow, etc.

Test application #1: fruit classifier[edit | edit source]

This application implements a classifier like the one described MISC-TN-011:_Running_an_Azure-generated_TensorFlow_Lite_model_on_Mito8M_SoM_using_NXP_eIQ here. There is one notable difference, however. The model

TBD