학술논문

Supporting Text Entry in Virtual Reality with Large Language Models
Document Type
Conference
Source
2024 IEEE Conference Virtual Reality and 3D User Interfaces (VR) VR Virtual Reality and 3D User Interfaces (VR), 2024 IEEE Conference. :524-534 Mar, 2024
Subject
Computing and Processing
Three-dimensional displays
Layout
Prototypes
Collaboration
Manuals
Virtual reality
Organizations
Human-centered computing
Human computer interaction (HCI)
Interaction paradigms
Virtual Reality
Interaction techniques
Text input
Language
ISSN
2642-5254
Abstract
Text entry in virtual reality (VR) often faces challenges in terms of efficiency and task loads. Prior research has explored various solutions, including specialized keyboard layouts, tracked physical devices, and hands-free interaction. Yet, these efforts often fall short of replicating the efficiency of real-world text entry, or introduce additional spatial and device constraints. This study leverages the extensive capabilities of large language models (LLMs) in context perception and text prediction to enhance text entry efficiency by reducing users’ manual keystrokes. Three LLM-assisted text entry methods - Simplified Spelling, Content Prediction, and Keyword-to-Sentence Generation - are introduced, aligning with user cognition and the contextual predictability of English text at word, grammatical structure, and sentence levels. Through user experiments encompassing various text entry tasks on an Oculus-based VR prototype, these methods demonstrate a 16.4%, 49.9%, 43.7% reduction in manual keystrokes, translating to efficiency gains of 21.4%,74.0%, 76.3%, respectively. Importantly, these methods do not increase manual corrections compared to manual typing, while significantly reducing physical, mental, and temporal loads and enhancing overall usability. Long-term observations further reveal users’ strategies for using these LLM-assisted methods, showing that users’ proficiency with the methods can reinforce their positive effects on text entry efficiency.