Abstract
XR devices running chat-bots powered by Large Language Models (LLMs) have the to become always-on agents that enable much better productivity scenarios. Current screen based chat-bots do not take advantage of the the full-suite of natural inputs available in XR, including inward facing sensor data, instead they over-rely on explicit voice or text prompts, sometimes paired with multi-modal data dropped as part of the query. We propose a solution that leverages an attention framework that derives context implicitly from user actions, eye-gaze, and contextual memory within the XR environment. Our work minimizes the need for engineered explicit prompts, fostering grounded and intuitive interactions that glean user insights for the chat-bot.
| Original language | English |
|---|---|
| Title of host publication | Proceedings - 2025 IEEE Conference on Virtual Reality and 3D User Interfaces, VR 2025 |
| Publisher | Institute of Electrical and Electronics Engineers Inc. |
| Pages | 708-717 |
| Number of pages | 10 |
| ISBN (Electronic) | 9798331536459 |
| DOIs | |
| Publication status | Published - 2025 |
| Event | 32nd IEEE Conference on Virtual Reality and 3D User Interfaces, VR 2025 - Saint-Malo, France Duration: 8-Mar-2025 → 12-Mar-2025 |
Conference
| Conference | 32nd IEEE Conference on Virtual Reality and 3D User Interfaces, VR 2025 |
|---|---|
| Country/Territory | France |
| City | Saint-Malo |
| Period | 08/03/2025 → 12/03/2025 |
Keywords
- AI Agents
- AI input
- Chatbots
- Multi-window
- XR productivity