dc.description.abstract | With the rapid advancement of technology, augmented reality (AR), virtual reality
(VR), and mixed reality (MR) have become highly regarded fields. However,
these technologies still face several challenges in remote rendering applications,
including the integration of AI-driven voice interactions. For instance,
a single server often proves inadequate in dynamic environments due to its
inability to meet the demands of high-complexity graphics rendering and realtime
AI voice processing. The limited CPU and GPU resources of a server
can quickly become overloaded as user load increases, leading to performance
degradation, increased latency, and potential system crashes.To address these
challenges, this paper proposes a multi-server architecture designed to effectively
handle AI voice interactions within AR, VR, and MR environments. By
distributing workload across multiple servers and optimizing resource allocation,
this architecture enhances overall system performance and improves user
experience. Furthermore, server-side object streaming camera design and local
positioning assistance mitigate localization errors induced by latency, surpassing
existing methods. | en_US |