Ph.D. Student in Computer Science
Texas A&M University

About Me

Human-Computer Interaction | Novel Sensing | Digital Fabrication | AI

I am a third year Ph.D. student at the HCIED (HCI Engineering & Design) Lab. At a high level, my big research direction is to enable everyday objects to sense, respond, and adapt to human activities. Within this vision, my current interested research areas overlapping digital fabrication and materials, novel sensing, and advanced cutting-edge technologies, like mixed reality, AI, and Computer Vision.

News

  • Oct. 2025 - I passed my PhD prelim exam!
  • Aug. 2025 - One paper accepted at UIST'25.
  • Jan. 2025 - My first first-authored long paper accepted at CHI'25!
  • Jan. 2024 - One paper accepted at CHI'24.
  • Jan. 2023 - Started my PhD life at Texas A&M!
  • May. 2022 - Presented my LBW paper virtually at CHI'22.
  • Feb. 2022 - My first LBW paper accepted at CHI'22!
  • Feb. 2022 - My first time entering industry as a full-time software engineer.
  • Dec. 2021 - I graduated from Texas A&M with B.S. Computer Science!

Publications


The gif showing different LuxAct real life applications.

LuxAct: Enhance Everyday Objects for Visual Sensing with Interaction-Powered Illumination (UIST'25)

Xiaoying Yang, Qian Lu, Jeeun Kim, Yang Zhang

a self-powered, ultra-low-cost system that lets everyday objects encode interaction and sensor data into RGB light patterns that AR headsets can read, enabling rich, ubiquitous sensing without embedded electronics.

DOI |

The gif showing different LumosX samples compared under regular camera and app view side by side.

LumosX: 3D Printed Anisotropic Light-Transfer (CHI'25)

Qian Lu, Xiaoying Yang, Xue Wang, Jacob Sayono, Yang Zhang, Jeeeun Kim

A set of novel techniques for encoding and decoding information through light intensity changes using 3D-printed optical anisotropic properties.

DOI |

The figure shows a flowchart from left to right. On the left, there is a 5 by 6 grid of images with a bounding box around an item in each image, all from AccessDB and AccessReal. The grid then points to a network icon labeled Inaccessibility Detector, which points to a mobile phone screen that has an image of a bathroom, with bounding boxes around several items in the photo. One of the bounding boxes around the outlet on the wall has an arrow pointing to another diagram that has assistive designs from AccessMeta sorted in three rows: Actuation, Indication, and Constraint from top to bottom.

AccessLens: Auto-detecting Inaccessibility of Everyday Objects (CHI'24)

Nahyun Kwon, Qian Lu, Muhammad Hasham Qazi, Joanne Liu, Changhoon Oh, Shu Kong, Jeeeun Kim

An end-to-end system designed to identify inaccessible interfaces in daily objects, and recommend 3D-printable augmentations for accessibility enhancement.

DOI |

The image shows two configurations of Programmable Filament. The upper one is series configuration, and the lower one is parallel configuration. Both configurations use PLA and TPU as example.

User-Centered Property Adjustment with Programmable Filament (CHI'22 LBW)

Qian Lu, Aryabhat Darnal, Haruki Takahashi, Anastasia Hanifah Muliana, Jeeeun Kim

We discovered that Programmable Filament, a novel technique to enable end-users with low-cost Fused Deposition Modeling (FDM) 3D printers to equip multi-material printing capabilities at low investment, can be applied to mix mechanical properties of two different materials (e.g., tensility) that may affect user’s comfort in 3D printed objects (e.g., personal optimal softness in sports gear) to meet individual needs.

DOI