MODI: Mobile Deep Inference Made Efficient by Edge Computing


Abstract

In this paper, we propose a novel mobile deep inference platform, MODI, that delivers good inference performance. MODI improves deep learning powered mobile applications performance with optimizations in three complementary aspects. First, MODI provides a number of models and dynamically selects the best one during runtime. Second, MODI extends the set of models each mobile application can use by storing high quality models at the edge servers. Third, MODI manages a centralized model repository and periodically updates models at edge locations, ensuring up-to-date models for mobile applications without incurring high network latency. Our evaluation demonstrates the feasibility of trading off inference accuracy for improved inference speed, as well as the acceptable performance of edge-based inference.


Ogden2018 PDF


BibTeX

@InProceedings{Ogden2018, author = {Samuel S. Ogden and Tian Guo}, booktitle = {{USENIX} Workshop on Hot Topics in Edge Computing (HotEdge 18)}, title = {{MODI}: Mobile Deep Inference Made Efficient by Edge Computing}, year = {2018}, address = {Boston, MA}, month = jul, publisher = {{USENIX} Association}, abstract = {In this paper, we propose a novel mobile deep inference platform, MODI, that delivers good inference performance. MODI improves deep learning powered mobile applications performance with optimizations in three complementary aspects. First, MODI provides a number of models and dynamically selects the best one during runtime. Second, MODI extends the set of models each mobile application can use by storing high quality models at the edge servers. Third, MODI manages a centralized model repository and periodically updates models at edge locations, ensuring up-to-date models for mobile applications without incurring high network latency. Our evaluation demonstrates the feasibility of trading off inference accuracy for improved inference speed, as well as the acceptable performance of edge-based inference.}, url = {https://www.usenix.org/conference/hotedge18/presentation/ogden}, }