I. Introduction
With the fast development of mobile edge computing (MEC), many interactive mobile augmented reality (MAR) applications, such as Pokemon Go, Connected Lens, and Just a Line, have exploded in popularity among mobile end users. In these applications, virtual objects are overlaid on the frames captured by the users and remain fixed relative to the real world [1]. This demands that the view is well understood, which can be enabled with deep neural network (NN)-based video analytics techniques. Since mobile devices are often resource-limited to execute large networks, the users can offload the workloads to the edge servers [5], [6]. In this way, they can achieve much lower latency for a better experience. Moreover, users can share their virtual objects with each other to enjoy the interaction.