Jing Qian’s research while affiliated with New York University and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (9)


Satori 悟り: Towards Proactive AR Assistant with Belief-Desire-Intention User Modeling
  • Conference Paper

April 2025

·

6 Reads

·

1 Citation

Chenyi Li

·

Guande Wu

·

·

[...]

·

Jing Qian

Design and Implementation of the Transparent, Interpretable, and Multimodal (TIM) AR Personal Assistant

April 2025

·

13 Reads

The concept of an AI assistant for task guidance is rapidly shifting from a science fiction staple to an impending reality. Such a system is inherently complex, requiring models for perceptual grounding, attention, and reasoning, an intuitive interface that adapts to the performer's needs, and the orchestration of data streams from many sensors. Moreover, all data acquired by the system must be readily available for post-hoc analysis to enable developers to understand performer behavior and quickly detect failures. We introduce TIM, the first end-to-end AI-enabled task guidance system in augmented reality which is capable of detecting both the user and scene as well as providing adaptable, just-in-time feedback. We discuss the system challenges and propose design solutions. We also demonstrate how TIM adapts to domain applications with varying needs, highlighting how the system components can be customized for each scenario.


Design and Implementation of the Transparent, Interpretable, and Multimodal (TIM) AR Personal Assistant

March 2025

·

16 Reads

IEEE Computer Graphics and Applications

The concept of an AI assistant for task guidance is rapidly shifting from a science fiction staple to an impending reality. Such a system is inherently complex, requiring models for perceptual grounding, attention, and reasoning, an intuitive interface that adapts to the performer's needs, and the orchestration of data streams from many sensors. Moreover, all data acquired by the system must be readily available for post-hoc analysis to enable developers to understand performer behavior and quickly detect failures. We introduce TIM, the first end-to-end AI-enabled task guidance system in augmented reality which is capable of detecting both the user and scene as well as providing adaptable, just-in-time feedback. We discuss the system challenges and propose design solutions. We also demonstrate how TIM adapts to domain applications with varying needs, highlighting how the system components can be customized for each scenario.


Figure 2: The front and backend systems. A change in arrow color represents a change in information.
Figure 3: Cognitive classifier: fNIRS data is preprocessed using wavelet filters over a sliding window then the 3 cognitive facets are classified.
AdaptiveCoPilot: Design and Testing of a NeuroAdaptive LLM Cockpit Guidance System in both Novice and Expert Pilots
  • Preprint
  • File available

January 2025

·

123 Reads

Pilots operating modern cockpits often face high cognitive demands due to complex interfaces and multitasking requirements, which can lead to overload and decreased performance. This study introduces AdaptiveCoPilot, a neuroadaptive guidance system that adapts visual, auditory, and textual cues in real time based on the pilot's cognitive workload, measured via functional Near-Infrared Spectroscopy (fNIRS). A formative study with expert pilots (N=3) identified adaptive rules for modality switching and information load adjustments during preflight tasks. These insights informed the design of AdaptiveCoPilot, which integrates cognitive state assessments, behavioral data, and adaptive strategies within a context-aware Large Language Model (LLM). The system was evaluated in a virtual reality (VR) simulated cockpit with licensed pilots (N=8), comparing its performance against baseline and random feedback conditions. The results indicate that the pilots using AdaptiveCoPilot exhibited higher rates of optimal cognitive load states on the facets of working memory and perception, along with reduced task completion times. Based on the formative study, experimental findings, qualitative interviews, we propose a set of strategies for future development of neuroadaptive pilot guidance systems and highlight the potential of neuroadaptive systems to enhance pilot performance and safety in aviation environments.

Download

Satori: Towards Proactive AR Assistant with Belief-Desire-Intention User Modeling

October 2024

·

26 Reads

Augmented Reality assistance are increasingly popular for supporting users with tasks like assembly and cooking. However, current practice typically provide reactive responses initialized from user requests, lacking consideration of rich contextual and user-specific information. To address this limitation, we propose a novel AR assistance system, Satori, that models both user states and environmental contexts to deliver proactive guidance. Our system combines the Belief-Desire-Intention (BDI) model with a state-of-the-art multi-modal large language model (LLM) to infer contextually appropriate guidance. The design is informed by two formative studies involving twelve experts. A sixteen within-subject study find that Satori achieves performance comparable to an designer-created Wizard-of-Oz (WoZ) system without relying on manual configurations or heuristics, thereby enhancing generalizability, reusability and opening up new possibilities for AR assistance.



ARGUS: Visualization of AI-Assisted Task Guidance in AR

November 2023

·

93 Reads

·

29 Citations

IEEE Transactions on Visualization and Computer Graphics

The concept of augmented reality (AR) assistants has captured the human imagination for decades, becoming a staple of modern science fiction. To pursue this goal, it is necessary to develop artificial intelligence (AI)-based methods that simultaneously perceive the 3D environment, reason about physical tasks, and model the performer, all in real-time. Within this framework, a wide variety of sensors are needed to generate data across different modalities, such as audio, video, depth, speech, and time-of-flight. The required sensors are typically part of the AR headset, providing performer sensing and interaction through visual, audio, and haptic feedback. AI assistants not only record the performer as they perform activities, but also require machine learning (ML) models to understand and assist the performer as they interact with the physical world. Therefore, developing such assistants is a challenging task. We propose ARGUS, a visual analytics system to support the development of intelligent AR assistants. Our system was designed as part of a multi-year-long collaboration between visualization researchers and ML and AR experts. This co-design process has led to advances in the visualization of ML in AR. Our system allows for online visualization of object, action, and step detection as well as offline analysis of previously recorded AR sessions. It visualizes not only the multimodal sensor data streams but also the output of the ML models. This allows developers to gain insights into the performer activities as well as the ML models, helping them troubleshoot, improve, and fine-tune the components of the AR assistant.



ARGUS: Visualization of AI-Assisted Task Guidance in AR

August 2023

·

70 Reads

·

3 Citations

The concept of augmented reality (AR) assistants has captured the human imagination for decades, becoming a staple of modern science fiction. To pursue this goal, it is necessary to develop artificial intelligence (AI)-based methods that simultaneously perceive the 3D environment, reason about physical tasks, and model the performer, all in real-time. Within this framework, a wide variety of sensors are needed to generate data across different modalities, such as audio, video, depth, speech, and time-of-flight. The required sensors are typically part of the AR headset, providing performer sensing and interaction through visual, audio, and haptic feedback. AI assistants not only record the performer as they perform activities, but also require machine learning (ML) models to understand and assist the performer as they interact with the physical world. Therefore, developing such assistants is a challenging task. We propose ARGUS, a visual analytics system to support the development of intelligent AR assistants. Our system was designed as part of a multi year-long collaboration between visualization researchers and ML and AR experts. This co-design process has led to advances in the visualization of ML in AR. Our system allows for online visualization of object, action, and step detection as well as offline analysis of previously recorded AR sessions. It visualizes not only the multimodal sensor data streams but also the output of the ML models. This allows developers to gain insights into the performer activities as well as the ML models, helping them troubleshoot, improve, and fine tune the components of the AR assistant.

Citations (3)


... Wu et al. [46] proposed an automatic text-simplification tool to assist users when navigating virtual environements. The authors evaluated how well ARTiST guides users through physical AR tasks compared to a baseline system. ...

Reference:

PLanet: Formalizing Experimental Design
ARTiST: Automated Text Simplification for Task Guidance in Augmented Reality
  • Citing Conference Paper
  • May 2024

... Current robots often struggle to effectively use chemical cues [25]. New benchmarks are needed to test an AI's proficiency in pinpointing odour sources, such as a gas leak in a building [24,55], tracking dynamic scent plumes [137], or navigating using sparse olfactory landmarks [41] (c) Reasoning tasks (akin to video question answering [68,160,115,114], multimodal event understanding [58,113], physical task guidance and anomaly detection [126,27,163,95] and robot manipulation [157,87]) where focus on answering complex questions or generating summaries about dynamic events and activities by integrating information from olfactory, visual (video), and auditory streams. For example, "Based on the sight of smoke, the sound of a crackling fire, and the smell of burning wood, what is likely happening and where?" ...

ARGUS: Visualization of AI-Assisted Task Guidance in AR
  • Citing Article
  • November 2023

IEEE Transactions on Visualization and Computer Graphics

... These include voice assistants like Siri, Alexa, or Google, and some even support visual modality (e.g., [1,30]), enabling rich inputs and outputs. Similar to science fiction, advanced digital assistants can practically aid users in performing both familiar and new tasks, reduce task load and errors, and enhance task performance [15]. ...

ARGUS: Visualization of AI-Assisted Task Guidance in AR
  • Citing Preprint
  • August 2023