My research focuses on generative models applied across a range of computer vision tasks, with particular focus on 3D human motion understanding, realistic animation generation, and clinical gait analysis.
PickStyle is a diffusion-based video style transfer framework that preserves video context while applying a target visual style. It uses low-rank style adapters and synthetic clip augmentation from paired images for training, and introduces Context-Style Classifier-Free Guidance (CS-CFG) to independently control content and style, achieving temporally consistent and style-faithful video results.
CARE-PD is the largest publicly available archive of 3D mesh gait data for Parkinson’s Disease, collected across 9 cohorts from 8 clinical centers. It provides standardized, anonymized SMPL representations and benchmark protocols for clinical motion analysis on PD.
GAITGen is a generative framework that synthesizes realistic gait sequences conditioned on Parkinson’s severity. Using a Conditional Residual VQ-VAE and tailored Transformers, it disentangles motion and pathology features to produce clinically meaningful gait data. GAITGen enhances dataset diversity and improves performance in parkinsonian gait analysis tasks.
A unified end-to-end model that predicts both human global motion (trajectory) and detailed body pose jointly, using social and scene context to improve forecasting.
Proposes a component-based video content representation for human action recognition, decomposing videos into meaningful parts to improve classification in complex scenes.