About

I am Yuqiang Yang, a Research&Development Engineer at Embodied AI Center, Shanghai AI Laboratory. Working with Dr. Tai Wang and Dr. Jiangmiao Pang, my research interests are to combine LLM with robotic model-based optimization to enable various robots, such as mobile manipulator, autonomous car, humanoid and multicopter, to perform tasks autonomously and efficiently in unstructured environment.

Specifically, we study post-training the VLM-based System2 model and diffusion-based obstacle avoidance System1 model to enable different embodiments (Unitree G1, Go2, Turtlebot, Galaxea, etc) to finish VLN tasks efficently and smoothly. Besides, we have explored and realized the possibility of allowing the mobile manipulators to efficiently pick and place in a wholebody manner while avoidance collision smoothly in cluttered dynamic environment full of chairs, tables and shelves. I have full-stack technical expertise across core domains including high-quality data generation, efficient model evaluation, sim2real transfer, robust robotics localization, high-dimensional motion planning and high-precision optimal control.

Our team is dedecated to building Embodied AGI systems and empowering academia and industry through open-source initiatives. We have contributed many works on github. If you are interested to join us, feel free to contact us.

Education and Training

South China University of Technology
Master, Robotics
Supervisor: Prof. Chenguang Yang
GPA: 3.82/4.0 (ranked first)
Sep. 2022 - Present





South China University of Technology
Bachelor of Engineering, Automation
School of Automation Science and Engineering
GPA: 3.94/4.0 (ranked first)
Sep. 2018 - Jun. 2022





FastLab of Zhejiang University
Visiting student
Wholebody planning and control for multicopter
Supervisor: Prof. Fei Gao
Oct. 2023 - Nov. 2023




Research Experiences

streamvln

StreamVLN: Streaming Vision-and-Language Navigation via SlowFast Context Modeling

2025 ArXiv preprint

[Project Page] [Paper] [Code] [Zhihu]

We propose StreamVLN, a streaming VLN framework that employs a hybrid slow-fast context modeling strategy to support multi-modal reasoning over interleaved vision, language and action inputs. StreamVLN can understand complex human instructions and finish VLN task in various indoor/outdoor scenarios.

navdp

NavDP: Learning Sim-to-Real Navigation Diffusion Policy with Privileged Information Guidance

2025 ArXiv preprint

[Project Page] [Paper] [Code] [Zhihu]

We propose NavDP, an end-to-end framework trained solely in simulation that combines diffusion-based trajectory generation and a critic function for trajectory selection (conditioned on local observation tokens from a shared policy transformer); it can zero-shot transfer to different robot embodiments in diverse real-world environments.

ESDF Map Planning Trajectory
Real-World Experiment

RAMPAGE: Towards Whole-body, Real-Time and Agile Motion Planning in Dynamic Cluttered Environments for Mobile Manipulators

[Video] [PDF]

2024 IEEE Transaction on Industrial Electronics

We proposed a hierarchical topology-guided search and AL-DDP-based optimization to solve whole-body kinodynamic planning in dynamic environments, and achieved real-time planning (≈30ms) and ≈80% success rate with accurate collision detection via ESDF map and sphere decomposition.

Door Opening Demo

Learn to Coordinate: a Whole-Body Learning from Demonstration Framework for Differential Drive Mobile Manipulators

[Video1, Video2] [PDF]

2023 IEEE Conference on Systems, Man, and Cybernetics

We developed a Gaussian Process-based learning framework with WLN inverse kinematics for few-shot whole-body skill learning, which enabled coordinated door opening with disturbance rejection and simplified human guidance via admittance control.

Project Experiences

H1 in Hospital H1 in Office H1 Outdoors H1 on Stairs

Locomotion in complex terrain through reinforcement learning in Isaac lab

[Video]

We used PPO algorithm with a fine-tuned reward function to train Unitree H1’s locomotion via curriculum learning, and realized robust Sim-to-Sim transfer of RL policies to diverse photo-realistic environments in Nvidia Isaac Sim.

Multicopter in Forest

Risk-aware contingency motion planning under uncertainties for Automated Valet Parking(AVP)

DJI Automotive

We proposed Voronoi-based safe corridors and SplineGrid optimization for lateral bypass plus iLQR for risk-aware longitudinal speed, and handled prediction multimodality via tree-branch iLQR to ensure safe trajectory under perception/prediction uncertainties.

Multicopter in Forest

Low-cost and efficient location, mapping, planning and control for multicopter in embedded system

[Video1, Video2, Video3]

Application Innovate Laboratory, Huawei; FastLab, Zhejiang University | 2023.2 - 2024.3

We enhanced VINS-Fusion with learning-based features and QR optimization for better accuracy/robustness on low-performance chips, and implemented fast OGM updating (incremental inflation) and robot-centric ESDF for efficient MPCC-based collision-aware control.

SE3 Narrow Gap Crossing

SE3 planning and control for multicopter to cross narrow gap

[Video]

FastLab, Zhejiang University | 2023.10 - 2023.12

We built safe flight corridors (SFC) and used MINCO for spatial-temporal trajectory optimization via L-BFGS, and achieved accurate narrow gap crossing by calibrating thrust mapping and tuning controllers for large-attitude tracking.

Pedestrian Following Demo

Pedestrian following and collision avoidance with spatial-temporal optimization for differential car

[Video]

Application Innovate Laboratory, Huawei; FastLab, Zhejiang University | 2023.2 - 2023.11

We developed multi-level hybrid A* for EKF-predicted pedestrian following plus MINCO optimization for smooth trajectories, and ensured collision avoidance via lidar filtering and MPC control to handle kinematic constraints and communication delay.

Self-balanced Racing Car

Self-balanced racing car with wireless charging capability

[Video]

School of Automation Science and Engineering, SCUT | 2020.01 - 2020.08

We designed adaptive PD wireless charging (≈30W) for super-capacitors and tuned cascade controllers for stable track navigation, which helped win 5th national place with a 23.8s race time and successful passage through circles, slopes, and crossroads.

Wholebody Pick-and-Place

Wholebody pick-and-place for mobile manipulator

[Video]

Application Innovate Laboratory, Huawei | 2022.10 - 2022.11

We trained GGCNN for 6-D object perception and used OSQP to solve dynamic-weighted QP for whole-body coordination, which enabled smooth pick-and-place by balancing manipulability, energy, and trajectory tracking performance.

Honor & Awards

Internship

DJI Automotive
Intern, PnC
Decision and motion planning for autonomous vehicles
Supervisor: Dr. Yifan Tang, Dr. Zhepei Wang
Apr. 2024 - Nov. 2024





Huawei Technologies Co.Ltd
Intern, Application Innovation Laboratorys
Planning and Control for various robots
Supervisor: Dr. Chen Chen, Dr. Zehui Meng
June. 2022 - Apr. 2024





China-Singapore International Joint Research Institute
Intern, Robot Perception and Computer Vision Group
Multi-sensor calibration and 3D detection
Supervisor: Dr. Mingxing Wen
Jan. 2021 - Mar. 2021




Skills

Programming:

Python, MATLAB, C/C++, PyTorch, Pybullet, Airsim, Embedded System

Robotics:

Wholebody control, Peception and mapping, Convex Optimization, Admittance/Impedance Control, Gravity Compensation, teleoperation

Embodied AI:

Data preparation, Model training and evaluation, Sim2Real Transfer

Hardware Experience:

Unitree G1, Unitree Go2, Turtlebot4, Galaxea R1, Multicopter, Diablo, Franka, UR, Mobile Manipulator, Robotiq 2F85, Vicon, Touch X, ATI sensors, STM32