ARM’s Zach Shelby introduced the use of microcontrollers for machine learning and artificial intelligence at the ECF19 event in Helsinki on last Friday. The talk showed that that artificial intelligence and machine learning can be applied to small embedded devices in addition to the cloud-based model. In particular, artificial intelligence is well suited to the devices of the Internet of Things. The use of machine learning in IoT is also sensible from an energy efficiency point of view if unnecessary power-consuming communication can be avoided (for example local keyword detection before sending voice data to cloud more more detailed analysis).
According to Shelby , we are now moving to a third wave of IoT that comes with comprehensive equipment security and voice control. In this model, machine learning techniques are one new application that can be added to previous work done on IoT.
In order to successfully use machine learning in small embedded devices, the problem to be solved is that it has reasonably little incoming information and a very limited number of possible outcomes. ARM Cortex M4 processor equipped with a DSP unit is powerful enough for simple hand writing decoding or detecting few spoken words with machine learning model. In examples the used machine learning models needed less than 100 kilobytes of memory.
The presentation can be now viewed on YouTube:
Important tools and projects mentioned on the presentation:
uTensor (ARM MicroTensor)
Articles on presentation:
https://www.uusiteknologia.fi/2019/05/20/ecf19-koneoppiminen-mahtuu-mikro-ohjaimeen/
http://www.etn.fi/index.php/72-ecf/9495-koneoppiminen-mullistaa-sulautetun-tekniikan
420 Comments
Tomi Engdahl says:
Using an Edge Impulse tinyML model and an OLED library from Adafruit Industries, Atul Yadav developed a Nano 33 BLE Sense system that detects and displays when a myna bird sings.
Visual Indication for bird chirping (Mynah song detection)
https://m.youtube.com/watch?v=9ibp7H7qPnk
Tomi Engdahl says:
ML in Crop Quality and Environmental Tracking: A look at how many #environmental and #agricultural problems can be addressed within the same framework of audio analysis and environmental sensing through #machinelearning. https://bit.ly/3Ab89bf
A Continuously Sprouting Project: ML in Crop Quality and Environmental Tracking
https://www.sparkfun.com/news/3915?utm_content=171277991&utm_medium=social&utm_source=facebook&hss_channel=fbp-153488801415
A look at how many environmental and agricultural problems can be addressed with the same framework of audio analysis and environmental sensing through machine learning!
Tomi Engdahl says:
Use tinyML on the Nano 33 BLE Sense to classify different bird calls
https://blog.arduino.cc/2021/06/28/use-tinyml-on-the-nano-33-ble-sense-to-classify-different-bird-calls/
Tomi Engdahl says:
A FANtastic predictive maintenance project by Edge Impulse ambassador Andri Yadi! This system uses a tinyML model on an Arduino Nano 33 BLE Sense to classify fan operation and identify faults.
https://m.youtube.com/watch?v=GB6ArqH_eLo&feature=youtu.be
Tomi Engdahl says:
AIfES is a standalone AI framework that allows on-device training without a PC and works on almost any hardware. Even the 8-bit Arduino Uno!
AIfES is an AI/ML framework written in C for even the smallest microcontrollers
https://blog.arduino.cc/2021/07/06/aifes-is-an-ai-ml-framework-written-in-c-for-even-the-smallest-microcontrollers/
Tomi Engdahl says:
Recognising Bird Sounds With A Microcontroller
https://hackaday.com/2021/07/06/recognising-bird-sounds-with-a-microcontroller/
Machine learning is an incredible tool for conservation research, especially for scenarios like long term observation, and sifting through massive amounts of data. While the average Hackaday reader might not be able to take part in data gathering in an isolated wilderness somewhere, we are all surrounded by bird life. Using an Arduino Nano 33 BLE Sense and an online machine learning tool, [Errol Joshua] demonstrates how to set up an automated bird call classifier.
The Arduino Nano 33 BLE Sense is a fully featured little dev board that features the very capable NRF52840 microcontroller with Bluetooth Low Energy, and a variety of onboard sensors, including a microphone. Training a machine learning model might seem daunting to many people, but online services like Edge Impulse makes the process very beginner-friendly.
a massive online library of bird calls from all over the world is available on Xeno-Canto.
https://hackaday.io/project/180538-bird-sound-classifier-on-the-edge
Tomi Engdahl says:
This device uses machine learning to detect the ripening stages of various fruits and vegetables by spectral color.
Vegetables and Fruits Ripeness Detection by Color w/ TF © CC BY
https://create.arduino.cc/projecthub/kutluhan-aktar/vegetables-and-fruits-ripeness-detection-by-color-w-tf-041f92
Collate spectral color data of varying fruits and vegetables and interpret this data set with a neural network to predict ripening stages.
Tomi Engdahl says:
RECOGNISING BIRD SOUNDS WITH A MICROCONTROLLER
https://hackaday.com/2021/07/06/recognising-bird-sounds-with-a-microcontroller/
Machine learning is an incredible tool for conservation research, especially for scenarios like long term observation, and sifting through massive amounts of data. While the average Hackaday reader might not be able to take part in data gathering in an isolated wilderness somewhere, we are all surrounded by bird life. Using an Arduino Nano 33 BLE Sense and an online machine learning tool, a team made up of [Errol Joshua], [Ajith KJ], [Mahesh Nayak], and [Supriya Nickam] demonstrate how to set up an automated bird call classifier.
https://hackaday.io/project/180538-bird-sound-classifier-on-the-edge
Tomi Engdahl says:
‘Droop, There It Is!’ is a smart irrigation system that uses ML to visually diagnose drought stress
https://blog.arduino.cc/2021/07/13/droop-there-it-is-is-a-smart-irrigation-system-that-uses-ml-to-visually-diagnose-drought-stress/
Tomi Engdahl says:
Aksha is a Nano 33 BLE Sense-equipped pencil that uses tinyML to recognize different Hindi letters drawn in the air: https://medium.com/@naveenmanwani/aksha-an-arduino-based-ml-pencil-powered-by-tensorflow-lite-micro-e7ba854f42f3
Tomi Engdahl says:
Using a tinyML model on the Nano 33 BLE Sense, this device automatically detects when someone is snoring and begins to vibrate as an alert.
The Snoring Guardian listens while you sleep and vibrates when you start to snore
https://blog.arduino.cc/2021/07/20/the-snoring-guardian-listens-while-you-sleep-and-vibrates-when-you-start-to-snore/
Snoring is an annoying problem that affects nearly half of all adults and can cause others to lose sleep. Additionally, the ailment can be a symptom of a more serious underlying condition, so being able to know exactly when it occurs could be lifesaving. To help solve this issue, Naveen built the Snoring Guardian — a device that can automatically detect when someone is snoring and begin to vibrate as an alert.
Tomi Engdahl says:
Can tinyML enable you to “hear” the difference between pouring hot and cold water? Marcelo Rovai and Marco Zennaro used the Nano 33 BLE Sense and Edge Impulse to find out.
“Listening Temperature” with TinyML © GPL3+
https://create.arduino.cc/projecthub/mjrobot/listening-temperature-with-tinyml-7e1325
Can we “hear” a difference between pouring hot and cold water? Amazing proof-of-concept by a quick real deployment using Edge Impulse Studio
Tomi Engdahl says:
A̶I̶ A-Thigh. This wearable system counts your squats using a TensorFlow Lite model on the Nano 33 BLE Sense.
Squats Counter Using TensorFlow Lite and Tiny Motion Trainer © MIT
https://create.arduino.cc/projecthub/Manasmw333/squats-counter-using-tensorflow-lite-and-tiny-motion-trainer-4d9dcf
Squats counter that can detect squats by measuring the accelerometer readings and prints it on the Serial monitor.
Tomi Engdahl says:
https://blog.arduino.cc/2021/07/23/meet-your-new-table-tennis-coach-a-tinyml-powered-paddle/
Tomi Engdahl says:
ICYMI! During last week’s Arm #AITechTalk, Massimo Banzi showed off our machine learning-capable boards, some new tools and frameworks to support your project, and a few example tinyML applications to help get you started.
https://m.youtube.com/watch?v=mj2J445pdNc&feature=youtu.be
Tomi Engdahl says:
Is it possible for a neural network to “hear” the difference between pouring hot and cold water? Marcelo Rovai and Marco Zennaro decided to use a sound classification model on the Nano 33 BLE Sense to find out.
“Listening Temperature” with TinyML © GPL3+
https://create.arduino.cc/projecthub/mjrobot/listening-temperature-with-tinyml-7e1325
Can we “hear” a difference between pouring hot and cold water? Amazing proof-of-concept by a quick real deployment using Edge Impulse Studio
Tomi Engdahl says:
This tutorial shows how to train an artificial neural network using the AIfES ML framework on a PC and run it on your Arduino board afterwards.
How to Use AIfES on a PC for Training © GPL3+
https://create.arduino.cc/projecthub/aifes_team/how-to-use-aifes-on-a-pc-for-training-9ad5f8
This tutorial shows how to train an artificial neural network with AIfES on a PC and run it on your Arduino board afterwards.
Tomi Engdahl says:
BABL is a tinyML-powered baby monitor that uses a Nano 33 BLE Sense with Edge Impulse’s new EON Tuner to distinguish cries from other household sounds.
BABL x EON: A Baby Monitor Tuned by Edge Impulse! © MIT
https://create.arduino.cc/projecthub/ishotjr/babl-x-eon-a-baby-monitor-tuned-by-edge-impulse-b8aa30
Edge Impulse automatically improves model accuracy with one weird trick!
In the original BABL project, we achieved 86.3% accuracy without a ton of effort, but what if we could improve that number – and do so without a ton of manual tweaking and training? That’s the concept behind Edge Impulse’s new EON Tuner, which automatically generates the optimal model based on your target and dataset type. Here’s a look at how EON Tuner can dramatically improve accuracy in just a few clicks!
Tomi Engdahl says:
New to embedded machine learning? This tutorial walks you through the process of training a custom model with Edge Impulse and running it on the Portenta + Vision Shield.
https://docs.arduino.cc/tutorials/portenta-vision-shield/vs-openmv-ml
Tomi Engdahl says:
During the tinyML Foundation EMEA Technical Forum, University of Bologna PhD student Federica Zonzini explored how anomaly detection models on devices like the Nano 33 BLE Sense could be used for vibration-based structural health monitoring.
https://m.youtube.com/watch?v=RmUQzFTnGGY
Tomi Engdahl says:
https://blog.arduino.cc/2021/08/24/predicting-a-lithium-ion-batterys-life-cycle-with-tinyml/
Tomi Engdahl says:
Our friends at Edge Impulse recently announced official support for the Arduino Tiny Machine Learning Kit! This makes it easy to acquire images and other sensor data from your Nano 33 BLE Sense and OV7675 camera module, then build, train and deploy your ML model.
Announcing Support for the Arduino Tiny Machine Learning Kit
https://www.edgeimpulse.com/blog/announcing-support-for-the-arduino-tiny-machine-learning-kit
Tomi Engdahl says:
Keep tabs on the pH levels a hydroponic plant’s water supply with a Nano 33 BLE Sense and determine if it’s right using an Edge Impulse tinyML model.
Monitor the pH levels of a hydroponic plant’s water supply with Arduino and tinyML
https://blog.arduino.cc/2021/09/02/monitor-the-ph-levels-of-a-hydroponic-plants-water-supply-with-arduino-and-tinyml/
Tomi Engdahl says:
AIfES – Inference Tutorial © GPL3+
This tutorial shows the different ways how to perform an inference in AIfES®.
https://create.arduino.cc/projecthub/aifes_team/aifes-inference-tutorial-f44d96
Tomi Engdahl says:
Build And Code This AI Self-Driving Car For Less Than $110
https://www.iflscience.com/editors-blog/build-and-code-this-ai-selfdriving-car-for-less-than-110/
The car has a camera with built-in Bluetooth and Wi-Fi so you can connect it to other devices wirelessly. Wheelson also has an LCD display that allows you to see what the robot is viewing. It’s powered by four small electromotors and there’s a rechargeable Li-Po battery included.
Tomi Engdahl says:
Tauno Erik used tinyML on the Nano 33 BLE Sense to create a keyword spotting system that lights up the corresponding LED.
https://m.youtube.com/watch?v=GdRD4VHg54Q
Tomi Engdahl says:
https://hackaday.com/2021/09/06/smart-laser-cutter-ad-on-detects-material-optically/
Tomi Engdahl says:
TinyML on ajankohtainen aihe Telenor Research -tutkimusyksikön tutkimusohjelmassa. Käyttökohteita on loputtomasti aina autoista maatalouteen ja joukkoliikenteestä matkailuun. Mutta mitä TinyML tarkoittaa ja miksi sitä pidetään vallankumouksellisena? Lue lisää artikkelista!
Kuinka tehdä pienistä antureista älykkäitä? – Telenorin tutkimusyksikkö työstää seuraavaa vallankumouksellista IoT-innovaatiota
TinyML on ajankohtainen aihe Telenor Research -tutkimusyksikön tutkimusohjelmassa monestakin syystä. Se on täydellinen pari NB-IoT:lle, joka on IoT:lle optimoitu tietoverkko. Lisäksi sillä on loputon määrä käyttökohteita aina autoista maatalouteen ja joukkoliikenteestä matkailuun. Mitä TinyML siis tarkoittaa ja miksi sitä pidetään vallankumouksellisena?
https://www.dna.fi/yrityksille/blogi/-/blogs/kuinka-tehda-pienista-antureista-alykkaita-telenorin-tutkimusyksikko-tyostaa-seuraavaa-vallankumouksellista-iot-innovaatiota?utm_source=facebook&utm_medium=linkad&utm_content=ILTE-artikkeli-kuinka-tehda-pienista-antureista-alykkaita-telenorin-tutkimusyksikko-tyostaa-seuraavaa-vallankumouksellista-iot-innovaatiota&utm_campaign=H_MES_21-36-40_artikkelikampanja&fbclid=IwAR2GJX07CvOyxHsx9tez3vU-F24GL6dXVbkgUNdpM4sRHxeGzIDG3VlpWsE
Tomi Engdahl says:
A tinyML application with huge potential for EVs! Quickly predict a lithium-ion battery’s lifecycle using the Nano 33 BLE Sense and Edge Impulse.
Battery Life Cycle Predictor Powered by Edge Impulse © GPL3+
https://create.arduino.cc/projecthub/manivannan/battery-life-cycle-predictor-powered-by-edge-impulse-820ab8
A TinyML model to predict the Lithium Ion battery’s life cycle within shorter time using Edge Impulse.
Tomi Engdahl says:
Overengineering A Smart Doorbell
https://hackaday.com/2021/09/12/overengineering-a-smart-doorbell/
Fresh from the mediaeval splendour of the Belgian city of Gent, we bring you more from the Newline hacker conference organised by Hackerspace Gent. [Victor Sonck] works at the top of his house, and thus needed a doorbell notifier. His solution was unexpected, and as he admits over engineered, using machine learning on an audio stream from a microphone to detect the doorbell’s sound.
https://www.youtube.com/watch?v=GMLE4b0AKFE
Tomi Engdahl says:
https://www.dna.fi/yrityksille/blogi/-/blogs/kuinka-tehda-pienista-antureista-alykkaita-telenorin-tutkimusyksikko-tyostaa-seuraavaa-vallankumouksellista-iot-innovaatiota?utm_source=facebook&utm_medium=linkad&utm_content=ILTE-artikkeli-kuinka-tehda-pienista-antureista-alykkaita-telenorin-tutkimusyksikko-tyostaa-seuraavaa-vallankumouksellista-iot-innovaatiota&utm_campaign=H_MES_21-36-40_artikkelikampanja&fbclid=IwAR1Db6RZYMRzqiz6wDaQ6kfG8NG2gvlDTFVBUaNtwNWt0Aax37B8qtLxRTk
Tomi Engdahl says:
Jeremy Ellis wanted to see if the Arduino Portenta could recognize itself using tinyML.
And the answer is…
twitter.com/rocksetta/status/1438910453351022592
Tomi Engdahl says:
Use this tinyML tool to identify poison ivy on your next hike!
Poison ivy have you on edge? This is poison ivy at the Edge!
https://www.hackster.io/justinelutz/poison-ivy-have-you-on-edge-this-is-poison-ivy-at-the-edge-6d9975
Poison ivy causes painful rashes & is found in places with poor reception (forests). Use this edge ML tool to ID poison ivy on your hike!
Tomi Engdahl says:
TinyML Aerial Forest Fire Detection © MIT
Detecting forest fires with a long-range autonomous airplane equipped with satellite communication.
https://create.arduino.cc/projecthub/team-sol/tinyml-aerial-forest-fire-detection-78ec6b
Tomi Engdahl says:
Syntiant Partners with Edge Impulse for NDP101-Powered Arduino-Compatible TinyML Development Board
https://www.hackster.io/news/syntiant-partners-with-edge-impulse-for-ndp101-powered-arduino-compatible-tinyml-development-board-8116a480c347
Compact dev board can run speech applications at under 140 microwatts and sensor applications at under 100 microwatts, the company claims.
Tomi Engdahl says:
https://www.cnx-software.com/2021/08/29/esp32-s3-ai-capabilities-esp-dl-library/
Tomi Engdahl says:
This paper presents the successful first attempt at deploying federated transfer learning to a resource-constrained microcontroller, which in this case was a Nano 33 BLE Sense: arxiv.org/pdf/2110.01107.pdf
Tomi Engdahl says:
In his latest Eloquent Arduino post, Simone Salerno has put together a basic tutorial to help you create a gesture recognition system using Arduino.
https://eloquentarduino.github.io/2021/10/arduino-gesture-recognition-the-easy-way-with-machine-learning/
Tomi Engdahl says:
Performing gesture recognition on Arduino boards can be very straighforward and is composed by 2 steps: capture data to train a Machine Learning model on your PC, then export that model to plain C++ back into your board to predict the gestures. It only requires a very small amount of work on your side with the Eloquent Arduino software.
https://eloquentarduino.github.io/2021/10/arduino-gesture-recognition-the-easy-way-with-machine-learning/
Tomi Engdahl says:
TensorFlow Lite for MCUs is AI on the Edge
By Michael Parks for Mouser Electronics
https://www.mouser.com/empowering-innovation/ai?utm_source=endeavor&utm_medium=display&utm_campaign=ed-personifai-eit-ai-#article2-ai
TensorFlow, a Google-led effort, is a set of open-source software libraries that enable developers to easily integrate complex numerical computation algorithms and machine learning (ML) into their projects (Figure 1). According to Google, these libraries provide stable application programming interfaces for Python (Python 3.7+ across all platforms) and C. Also, they provide APIs without backward compatibility guarantees for C++, Go, Java and JavaScript. Additionally, an alpha release is available for Apple’s Swift language.
TensorFlow offers so-called end-to-end machine learning support for the development and utilization of deep neural networks (DNN). DNNs are an implementation of ML that are particularly adept at pattern recognition and object detection and classification. TensorFlow libraries support both phases of the machine-learning process, which are training and inferencing. The first is the training of deep neural networks that requires significant computing horsepower typically found in server-grade hardware and graphical processing units (GPUs). More recently application-specific integrated circuits known as Tensor Processing Unit (TPUs) have been developed to support the training efforts. The second phase, inferencing, is utilizing the trained DNNs in the real-world to respond to new inputs and make recommendations based on the analysis of those inputs against the trained models. This is the phase that should be of keen interest to embedded product developers.
The release of TensorFlow Lite for Microcontrollers (a subset of the TensorFlow libraries) is specifically geared for performing inferencing on memory-constrained devices typically found in most embedded systems applications. It does not allow you to train new networks. That still requires the higher-end hardware.
Tomi Engdahl says:
A winner of the TensorFlow Lite for Microcontrollers Challenge, Naveen Kumar’s device can be embedded in a pillow to listen throughout the night and vibrate to alert the sleeper when snoring is heard.
https://blog.arduino.cc/2021/10/20/the-snoring-guardian-listens-while-you-sleep-and-vibrates-when-you-start-to-snore/
Tomi Engdahl says:
Little Device, Big Data
https://www.hackster.io/news/little-device-big-data-d2660d2bc8da
Resource-constrained devices can continually learn from large datasets, while preserving data privacy, with TinyFedTL.
Tomi Engdahl says:
This system highlights the potential of tinyML for the performing arts.
Mapping Dance syncs movement and stage lighting using tinyML
https://blog.arduino.cc/2021/10/22/mapping-dance-syncs-movement-and-stage-lighting-using-tinyml/
Being able to add dynamic lighting and images that can synchronize with a dancer is important to many performances, which rely on both music and visual effects to create the show. Eduardo Padrón aimed to do exactly that by monitoring a performer’s moves with an accelerometer and triggering the appropriate AV experience based on the recognized movement.
Tomi Engdahl says:
Looking for a project to get started with machine learning on Arduino? Simone Salerno shows how to easily create a gesture recognition system using the Nano RP2040 Connect’s built-in accelerometer.
https://eloquentarduino.github.io/2021/10/arduino-gesture-recognition-the-easy-way-with-machine-learning/
Tomi Engdahl says:
RISC-V-Based VEGA Brings Continual Learning to TinyML with an Order of Magnitude Efficiency Gain
Based on the PULP Platform, this RISC-V chip could offer hundreds of thousands of hours of operation on a single battery charge.
https://www.hackster.io/news/risc-v-based-vega-brings-continual-learning-to-tinyml-with-an-order-of-magnitude-efficiency-gain-6189df9bb365
Tomi Engdahl says:
TinyML Dog Bark Stopper © GPL3+
An fun and simple project that uses TinyML to detect and respond to dog barks.
https://create.arduino.cc/projecthub/NathanielF/tinyml-dog-bark-stopper-77e436
Tomi Engdahl says:
Using Simone Salerno’s EloquentTinyML library, you only need a few lines of code to perform state-of-the-art person detection on the Portenta Vision Shield.
https://eloquentarduino.github.io/2021/11/3-lines-of-code-person-detection-on-arduino-portenta-and-esp32/
Person Detection on Arduino Portenta Vision Shield and ESP32 with Just 3 Lines of Code
10 NOVEMBER 2021
Have you ever wanted to perform person detection on Arduino or ESP32 boards? Found it difficult to custimize the sketch to suit your needs? Fear no more!
Person detection on Arduino and ESP32 microcontrollers doesn’t have to be difficult: with the right library, you only need 3 lines of code to perform state-of-the-art person detection.
EloquentTinyML is my library to run TensorFlow Neural Networks on Arduino boards without hassle.
EloquentTinyML both supports ARM Cortex-M (Arduino Portenta, Arduino Nano variants, STM32) boards and ESP32 boards.
I tested person detection on two different boards: Arduino Portenta + Vision Shield and M5Stack fisheye camera.
As promised, you will only need 3 lines of code to run person detection at the bare minimum:
Instantiate the detector
Configure the detector
Feed an image to the detector and get response back
The following is a working sketch that counts 60 lines of code including comments, debug messages and security checks!
Tomi Engdahl says:
OpenGL Machine Learning Runs On Low-End Hardware
https://hackaday.com/2021/11/13/opengl-machine-learning-runs-on-low-end-hardware/
If you’ve looked into GPU-accelerated machine learning projects, you’re certainly familiar with NVIDIA’s CUDA architecture. It also follows that you’ve checked the prices online, and know how expensive it can be to get a high-performance video card that supports this particular brand of parallel programming.
But what if you could run machine learning tasks on a GPU using nothing more exotic than OpenGL? That’s what [lnstadrum] has been working on for some time now, as it would allow devices as meager as the original Raspberry Pi Zero to run tasks like image classification far faster than they could using their CPU alone. The trick is to break down your computational task into something that can be performed using OpenGL shaders, which are generally meant to push video game graphics.
[lnstadrum] explains that OpenGL releases from the last decade or so actually include so-called compute shaders specifically for running arbitrary code. But unfortunately that’s not an option on boards like the Pi Zero, which only meets the OpenGL for Embedded Systems (GLES) 2.0 standard from 2007.
Constructing the neural net in such a way that it would be compatible with these more constrained platforms was much more difficult, but the end result has far more interesting applications to show for it. During tests, both the Raspberry Pi Zero and several older Android smartphones were able to run a pre-trained image classification model at a respectable rate.
Towards GPU-accelerated image classification on low-end hardware
https://medium.com/analytics-vidhya/towards-gpu-accelerated-image-classification-on-low-end-hardware-ec592e125ad9
Tomi Engdahl says:
Machine Learning Shushes Stressed Dogs
https://hackaday.com/2021/11/15/machine-learning-shushes-stressed-dogs/
He trained a TinyML neural net to detect when she barked and used and Arduino to play a sound byte to sooth her. The sound bytes in question are recordings of [Nathaniel]’s mom either praising or scolding [Clairette], and as you can see from the video below, they seem to work quite well. To train the network, [Nathaniel] worked with several datasets to avoid overfitting, including one he created himself using actual recordings of barks and ambient sounds within his own house. He used Eon Tuner, a tool by Edge Impulse, to help find the best model to use and perform the training. He uploaded the trained network to an Arduino Nano 33 BLE Sense running Mbed OS, and a second Arduino handled playing sound bytes via an Adafruit Music Maker Featherwing.
TinyML Dog Bark Stopper
An fun and simple project that uses TinyML to detect and respond to dog barks.
https://www.hackster.io/NathanielF/tinyml-dog-bark-stopper-77e436
Tomi Engdahl says:
Based on a Nano 33 BLE Sense, this prototype wristband doesn’t need a high-power radio and nearby computer to run a gesture-detecting CNN.
TinyML Gives an Arduino Nano 33 BLE Sense Wristband High-Speed, Low-Power Gesture Recognition
https://www.hackster.io/news/tinyml-gives-an-arduino-nano-33-ble-sense-wristband-high-speed-low-power-gesture-recognition-0e23b3ee1ebd
Built using low-cost parts, this prototype wristband doesn’t need a high-power radio and nearby computer to run a gesture-detecting CNN.