Reference 1
Reference 1 Eshratifar, Amir Erfan, Mohammad Saeed Abrishami, and Massoud Pedram. “JointDNN: an efficient training and inference engine for intelligent mobile cloud computing services.” IEEE Transactions on Mobile Computing (2019). ↩
Reference 1 Kim, Youngsok, Joonsung Kim, Dongju Chae, Daehyun Kim, and Jangwoo Kim. “μLayer: Low Latency On-Device Inference Using Cooperative Single-Layer Acceleration and Processor-Friendly Quantization.” In Proceedings of the Fourteenth EuroSys Conference 2019, pp. 1-15. 2019. ↩
Reference 1 Ogden, Samuel S., and Tian Guo. “{MODI}: Mobile Deep Inference Made Efficient by Edge Computing.” In {USENIX} Workshop on Hot Topics in Edge Computing (HotEdge 18). 2018. ↩
Reference 1 reference ↩
Reference 1 Cao, Qingqing, Niranjan Balasubramanian, and Aruna Balasubramanian. “MobiRNN: Efficient recurrent neural network execution on mobile GPU.” In Proceedings of the 1st International Workshop on Deep Learning for Mobile Systems and Applications, pp. 1-6. 2017. ↩
Reference 1 Liu, Sicong, Yingyan Lin, Zimu Zhou, Kaiming Nan, Hui Liu, and Junzhao Du. “On-demand deep model compression for mobile devices: A usage-driven model selection framework.” In Proceedings of the 16th Annual International Conference on Mobile Systems, Applications, and Services, pp. 389-400. 2018. ↩
Reference 1 reference ↩
Reference 1 Cho, Minsik, and Daniel Brand. “MEC: memory-efficient convolution for deep neural network.” In Proceedings of the 34th International Conference on Machine Learning-Volume 70, pp. 815-824. JMLR. org, 2017. ↩
Reference 1 Fang, Biyi, Xiao Zeng, and Mi Zhang. “Nestdnn: Resource-aware multi-tenant on-device deep learning for continuous mobile vision.” In Proceedings of the 24th Annual International Conference on Mobile Computing and Networking, pp. 115-127. 2018. ↩
Reference 1 reference ↩
Reference 1 Optimization of neural network inference to run on IoT/mobile devices. reference ↩
Reference 1 2016 Deepx Kit Reference 1 Lane, Nicholas D., Sourav Bhattacharya, Akhil Mathur, Claudio Forlivesi, and Fahim Kawsar. “DXTK: Enabling Resource-efficient Deep Learning on Mobile and Embedded Devices with the DeepX Toolkit.” In MobiCASE, pp. 98-107. 2016. ↩ Lane, Nicholas D., Sourav Bhattacharya, Petko Georgiev, Claudio Forlivesi, Lei Jiao, Lorena Qendro, and Fahim Kawsar. “Deepx: A software accelerator for low-power deep learning inference on mobile devices.
Reference 1 reference ↩
Reference1 AMC: leverage reinforcement learning to efficiently sample the design space. 4x FLOPs reduction, 2.7% better accuracy than hand-crafted model compression for VGG-16 on ImageNet. speedup 1.53x on GPU (Titan Xp) and 1.95x on Android phone (Google Pixel I), with negligible loss of accuracy. He, Yihui, Ji Lin, Zhijian Liu, Hanrui Wang, Li-Jia Li, and Song Han. “Amc: Automl for model compression and acceleration on mobile devices.” In Proceedings of the European Conference on Computer Vision (ECCV), pp.
If you could revise
the fundmental principles of
computer system design
to improve security...
... what would you change?