I. Introduction
The Metaverse is characterized by its immersive nature, allowing seamless integration between physical and virtual experiences, offering users the ability to live, work, and play in a digital environment [1]. For example, hands-on training with immersive simulations enables workers to refine their skills within a risk-free virtual realm [2]. Using virtual classrooms would allow students and teachers to interact in 3D virtual spaces for lectures, discussions, and collaborative activities [3]. Compared to the passive experience of live video streaming, the Metaverse is distinguished by its high level of interactivity. Real-time immersive Metaverse services place rigorous requirements on hardware and software to provide low delay and fast rendering performance. In terms of hardware, lightweight, low-power head-mounted displays (HMDs) and physical assistive devices are required to use high-resolution images for a long time without frequent charging or battery replacement [4]. In terms of software, rising delay adversely affects the user experience, potentially leading to dizziness and motion sickness due to sensory confusion [5]. Due to the limitations in computing power and battery capacity of HMD devices, balancing terminal energy consumption while ensuring immersive user experience and achieving high-quality, low-latency rendering on low-power devices has become an urgent challenge to address.