9. Inference Essentials
MB
Computing Time Memory Usage
Shorten the prediction time
is always welcome
Device memory is limited,
but deep learning model can
be huge
13. Nvidia TensorRT
Like a model compiler
Production Deep Learning with NVIDIA GPU Inference Engine, https://devblogs.nvidia.com/parallelforall/production-deep-learning-nvidia-gpu-inference-engine/
15. Quantization
How to Quantize Neural Networks with TensorFlow, https://github.com/tensorflow/tensorflow/blob/master/tensorflow/g3doc/how_tos/quantization/index.md
DNN is noise tolerable
FP16 to INT8
Hardware speedup
FPU to ALU
16. Inference Without DL Frameworks
Likely A compiler intermediate representation for image recognition and heterogeneous computing, http://liblikely.org/
26. Vision and Speech Limitations
Instead of
face identification
Keywords instead of NLP
FaceRecognition SpeechRecognition
27. Cloud Solution Drawbacks
CostConnectivity Privacy
Need to ensure bandwidth,
stability and latency are
good enough
Huge amount of
image transmission
You might want to keep
family information locally
56. Kinect v2
Update
USB Firmware
Open Source
Libraries
Fix data transmission issue libfreenect2 and pylibfreenect2
make enablement easier
MS Kinect v2 on Nvidia Jetson TX1, http://jetsonhacks.com/2016/07/11/ms-kinect-v2-nvidia-jetson-tx1/