Skip to content

Latest commit

 

History

History
82 lines (57 loc) · 5.36 KB

README.md

File metadata and controls

82 lines (57 loc) · 5.36 KB

Basic MNIST handwriting recognition with uTensor

uTensor reduces AI-inference cost significantly, bringing AI to Cortex M devices.

This tutorial discusses at a high level what TensorFlow graphs are and how to start using uTensor to build a handwriting recognition application. This involves training a fully-connected neural network on the MNIST dataset on a host machine, generating embedded code, and building an mbed application that classifies handwritten digits based on user input.

Fully connected 2 layer NN

Introduction

TensorFlow is an open-source software library for numerical computation using data flow graphs. Nodes in the graph represent mathematical operations, and the graph edges represent tensors, a type of multidimensional data array.

uTensor translates these TensorFlow graphs into C++ code, which you can run on embedded devices.

Running artificial intelligence on embedded systems involves 3 main steps.

  1. construct a machine learning model either offline on your local machine or in the cloud.
  2. feed this model into code generation tools. These tools output uTensor kernels, which you can run on your devices. These kernels are functions that output predictions based on their inputs.
  3. build embedded applications that call these functions to make informed decisions.

Requirements

Note for Windows users

NOTE: Windows is not currently supported; here are two workarounds:

  1. If you are running Windows then you must have both Python2 and Python3 installed. Tensorflow 1.2+ does not support Python 2.7 on Windows, and only officially supports Python 3.5 and 3.6. Meanwhile, mbed officially supports Python2 and has limited support for Python3. You can get the demo working by running the Ipython notebook with the Python3 kernel, and the mbed project build with Python2.

  2. Use the Cloud9 Environment

Build Instruction

  1. Import the project:
$ mbed import https://github.com/uTensor/utensor-mnist-demo

Train the model

Note

  • The repository contains reference model files. You may choose to skip to the Prepare the mbed project section.

  • If you prefer traditional Python scripts rather than notebooks, check out tensorflow-models/deep_mlp.py. This script contains all the same code as the notebook. The process of training and validating the model is exactly the same as in traditional machine learning workflows. This example trains a fully connected neural net with two hidden layers to recognize handwritten digits from the MNIST dataset, but you can apply the concepts to an application of your choice.

Jupyter Notebook (Optional)

In preparation for code generation, you must freeze the TensorFlow model. Freezing a model stores learned graph parameters in a protobuf file.

  1. Train the Tensorflow model
  2. Launch IPython jupyter notebook Launch the notebook
  3. Open the tensorflow-models/deep_mlp.ipynb
  4. Select Kernel/Restart & Run All. This will build a 2 layer NN then train, quantize, and save the model in tensorflow-models/mnist_model/deep_mlp.pb. Run the IPython notebook
  5. Go back to project root directory

Generate embedded C++ code

You can get the output node names from the IPython notebook. This will create a models directory. The models contains the embedded code interface for making inferences in your applications and relevant C++ source code associated with the learned weights and architecture of the neural net. Please refer to the uTensor-cli for installation and usage instructions.

# from project root, run:
$ utensor-cli convert tensorflow-models/mnist_model/deep_mlp.pb --output-nodes=y_pred

Prepare the mbed project

This example builds a handwriting recognition application using Mbed and the generated model, but you can apply these concepts to your own projects and platforms. This example uses the ST-Discovery-F413H because it has a touch screen and SD card built in, but you could just as easily build the application using plug-in components.

  1. Run mbed deploy, this fetches the necessary libraries like uTensor

  2. Build the mbed project:

    $ mbed compile -m DISCO_F413ZH -t GCC_ARM
    
  3. Finally flash your device by dragging and dropping the binary from BUILD/DISCO_F413ZH/GCC_ARM/utensor-mnist-demo.bin to your device.

Playing with the application

After drawing a number on the screen press the blue button to run inference, uTensor should output its prediction in the middle of the screen. Then press the reset button.
Whoops! Failed loading video

Note: The model used in training is very simple and has suboptimal accuracy in practice.