Research


Research Interests

My research mainly focuses on Physical AI, where AI enables physical systems—such as robotics, IoT, and embodied agents—to perceive, reason, and interact with the real world. Broadly speaking, I study how AI models enhance robotic perception and decision-making with multimodal sensor inputs (i.e. Multimodal AI, LLM, and VLA), how robots can personalize their behavior by understanding human states and interactions (i.e. Human-Centered Embodied AI), how foundation models can be efficiently deployed on edge devices to support real-time robotic perception and control (i.e. Edge Foundation Models for Robotics), and how robots generalize across embodiments and transfer knowledge from simulation to real-world tasks (i.e. Transfer Learning in Robotics). My ultimate vision is to create embodied AI that seamlessly integrates perception, reasoning, and action, enabling robots to assist humans across diverse environments with adaptive intelligence and edge efficiency.

Revolving these goals, my team mainly studies the following topics:

  • Multimodal AI, LLM, and VLA Models: Traditional AI models predominantly rely on a single modality, limiting their adaptability and robustness. Our research focuses on developing multimodal foundation models that integrate diverse sensor inputs—such as vision, language, and action signals—to enhance perception, reasoning, and decision-making. By leveraging large language models (LLMs) and Vision-Language-Action (VLA) models, we enable robots and AI systems to understand and interact with the world more effectively, improving their ability to generalize across diverse environments and tasks.

  • Human-Centered Embodied AI: Understanding human behavior is critical for advancing intelligent robotic systems. We leverage rich human perception data to improve human-robot interaction, prioritizing safety, adaptability, and intelligence. By modeling human behavior, pose dynamics, and cognitive states, we develop personalization strategies for embodied AI. These strategies ensure that robots can adapt their responses to individual users’ needs, preferences, and physical conditions, enhancing their autonomy and usability in real-world applications such as healthcare, assistive robotics, and industrial automation.

  • Transfer Learning in Robotics: Developing robots that can generalize across different embodiments and environments is a fundamental challenge in robotics. Our research explores cross-embodiment transfer learning, enabling knowledge transfer between different robotic platforms, and sim-to-real transfer, bridging the gap between simulated training environments and real-world deployment. By leveraging domain adaptation, reinforcement learning, and self-supervised techniques, we aim to enhance the adaptability of robotic systems, reducing the need for extensive real-world data collection while improving their performance in unstructured and dynamic environments.

  • Edge Foundation Models for Robotics: Deploying AI foundation models on resource-constrained edge devices is crucial for real-world robotic applications. Our research focuses on designing algorithms to minimize data and computational costs while preserving model accuracy. We develop efficient model compression, quantization, and optimization techniques to enable foundation models to run efficiently on edge hardware for robotic perception, planning, and control. Our ultimate goal is to create on-chip foundation models tailored for specific robotic tasks, ensuring scalable, low-latency, and energy-efficient AI inference.


Collaborations

I am interested in collaboration on the topics mentioned above. Apart from AI and IoT, I am also interested in AI-enabled interdisciplinary research, e.g., bioinformatics. Please drop me an email if you are interested in collaboration with a cool idea.