EmBARDiment: an Embodied AI Agent for Productivity in XR

  • Riccardo Bovo*
  • , Steven Abreu
  • , Karan Ahuja
  • , Eric J. Gonzalez
  • , Li Te Cheng
  • , Mar Gonzalez-Franco
  • *Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionAcademicpeer-review

1 Downloads (Pure)

Abstract

XR devices running chat-bots powered by Large Language Models (LLMs) have the to become always-on agents that enable much better productivity scenarios. Current screen based chat-bots do not take advantage of the the full-suite of natural inputs available in XR, including inward facing sensor data, instead they over-rely on explicit voice or text prompts, sometimes paired with multi-modal data dropped as part of the query. We propose a solution that leverages an attention framework that derives context implicitly from user actions, eye-gaze, and contextual memory within the XR environment. Our work minimizes the need for engineered explicit prompts, fostering grounded and intuitive interactions that glean user insights for the chat-bot.

Original languageEnglish
Title of host publicationProceedings - 2025 IEEE Conference on Virtual Reality and 3D User Interfaces, VR 2025
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages708-717
Number of pages10
ISBN (Electronic)9798331536459
DOIs
Publication statusPublished - 2025
Event32nd IEEE Conference on Virtual Reality and 3D User Interfaces, VR 2025 - Saint-Malo, France
Duration: 8-Mar-202512-Mar-2025

Conference

Conference32nd IEEE Conference on Virtual Reality and 3D User Interfaces, VR 2025
Country/TerritoryFrance
CitySaint-Malo
Period08/03/202512/03/2025

Keywords

  • AI Agents
  • AI input
  • Chatbots
  • Multi-window
  • XR productivity

Fingerprint

Dive into the research topics of 'EmBARDiment: an Embodied AI Agent for Productivity in XR'. Together they form a unique fingerprint.

Cite this