![8-Bit Quantization and TensorFlow Lite: Speeding up mobile inference with low precision | by Manas Sahni | Heartbeat 8-Bit Quantization and TensorFlow Lite: Speeding up mobile inference with low precision | by Manas Sahni | Heartbeat](https://miro.medium.com/max/1200/1*4_YzPSNvf_8rx8SvYvGTLA.jpeg)
8-Bit Quantization and TensorFlow Lite: Speeding up mobile inference with low precision | by Manas Sahni | Heartbeat
![How TensorFlow Lite Optimizes Neural Networks for Mobile Machine Learning | by Airen Surzyn | Heartbeat How TensorFlow Lite Optimizes Neural Networks for Mobile Machine Learning | by Airen Surzyn | Heartbeat](https://miro.medium.com/max/1400/1*MtXrCASxGrQtX2PPmhJcAw.png)
How TensorFlow Lite Optimizes Neural Networks for Mobile Machine Learning | by Airen Surzyn | Heartbeat
![8-Bit Quantization and TensorFlow Lite: Speeding up mobile inference with low precision | by Manas Sahni | Heartbeat 8-Bit Quantization and TensorFlow Lite: Speeding up mobile inference with low precision | by Manas Sahni | Heartbeat](https://miro.medium.com/max/1400/0*HjeBOLYllp9Q1pQj.png)
8-Bit Quantization and TensorFlow Lite: Speeding up mobile inference with low precision | by Manas Sahni | Heartbeat
![8-Bit Quantization and TensorFlow Lite: Speeding up mobile inference with low precision | by Manas Sahni | Heartbeat 8-Bit Quantization and TensorFlow Lite: Speeding up mobile inference with low precision | by Manas Sahni | Heartbeat](https://miro.medium.com/max/1400/1*xE-4bjdUJ9dHdgE7k74YZg.png)
8-Bit Quantization and TensorFlow Lite: Speeding up mobile inference with low precision | by Manas Sahni | Heartbeat
Hoi Lam 🇺🇦🇬🇧🇪🇺 on Twitter: "🚀New #TensorFlow Lite Android Support Library! Get more done with less boilerplate code for pre/post processing, quantization and label mapping: https://t.co/XyYJpZ9F4O Where are we going? 🎙️31 Oct
![Getting an error when creating the .tflite file · Issue #412 · tensorflow/model-optimization · GitHub Getting an error when creating the .tflite file · Issue #412 · tensorflow/model-optimization · GitHub](https://user-images.githubusercontent.com/38959661/83584586-74696700-a4fc-11ea-8072-3f095b53785e.png)
Getting an error when creating the .tflite file · Issue #412 · tensorflow/model-optimization · GitHub
![Quantized Conv2D op gives different result in TensorFlow and TFLite · Issue #38845 · tensorflow/tensorflow · GitHub Quantized Conv2D op gives different result in TensorFlow and TFLite · Issue #38845 · tensorflow/tensorflow · GitHub](https://user-images.githubusercontent.com/10414613/80136822-3ded1380-8570-11ea-8ef7-f1eee7211c0d.png)