Video Understanding for Activities of Daily Living
Temporal action detection, dense activity recognition, and long-video reasoning over unedited, real-world sequences.
Charlotte Vision Lab
UNC Charlotte
The Charlotte Vision Lab is a computer vision research group at UNC Charlotte focused on building systems that understand and act in the visual world. Our work spans video understanding, multimodal learning, robotic perception, generative modeling, and trustworthy machine vision. Our goal is to build trustworthy systems that can perceive, reason, and assist in complex real-world environments.
Research
Temporal action detection, dense activity recognition, and long-video reasoning over unedited, real-world sequences.
Visual question answering, domain adaptation, interpretable decision-making, and embodied reasoning.
3D scene understanding, controllable image generation, and uncertainty estimation in open-world settings.
Highlights
CVPR 2026
Temporal modeling for long untrimmed video understanding.
View paperCVPR 2026
A method to generate image descriptions in the style of a subject using attention sequences.
View paperCVPR 2025
A multimodal large language vision model for daily activities of living.
View paperPreprint
A depth-aware latent action framework for vision-language-action models in robot learning.
View paper