Boyang Wang
Hello!
This is Boyang Wang (ēåę“).
I am a first-year PhD student at the University of Virginia, supervised by Zezhou Cheng.
I graduated with a M.S. in Robotics and a B.S. in Computer Science from the University of Michigan (UMich).
At UMich, I worked with Prof. JJ Park, Prof. Nima Fazeli,
Prof. Hun-Seok Kim, and Prof. Chad Jenkins.
During the summer of 2025, I interned at Shanghai AI Lab, working closely with Xudong Xu on embodied AI department.
My research interests mainly lie in Generative Models (especially video) for General, Robotics, and Anime domains!
Email  Ā· 
Twitter  Ā· 
Github  Ā· 
Google Scholar
|
|
09/2025:
🎉
Frame In-N-Out is accepted by NeurIPS 2025! See you in San Diego!
08/2025:
🥳
Begin my journey as PhD student at the University of Virginia.
05/2025:
🎉
Frame In-N-Out: Unbounded Controllable Image-to-Video Generation is released on Arxiv.
02/2025:
🎉
One paper is accepted by CVPR 2025. See you in Nashville!
01/2025:
🎉
One paper is accepted by ICRA 2025. See you in Atlanta!
01/2025:
🎉
One paper is accepted by RA-L.
12/2024:
🎉
Graduated from University of Michigan as Master Degree.
08/2024:
🎉
Being selected as GSI for the EECS498 Computer Graphics and Generative Models.
07/2024:
🎉
This&That: Language-Gesture Controlled Video Generation for Robot Planning is released on Arxiv.
02/2024:
🎉
One paper is accepted by CVPR 2024.
10/2023:
🎉
One paper is accepted by WACV 2024.
|
Selected Publications
(* indicates equal contribution)
|
|
Frame In-N-Out: Unbounded Controllable Image-to-Video Generation
Boyang Wang,
Xuweiyi Chen,
Matheus Gadelha,
Zezhou Cheng
NeurIPS 2025
Paper
Ā· Website
First paper in Video Generation that studies Frame In and Frame Out cinematics effect, which is conditioned on First Frame Expanded Canvas, Text Prompt, Motion Trajectory and Identity Reference.
|
|
Sparse Image Sets Restoration with Multi-View Diffusion Model
Yucheng Mao *,
Boyang Wang *,
Nilesh Kulkarni,
Jeong Joon Park
CVPR 2025
Paper
Ā· Website
Reconstruct a 3D scene with degraded image sets by Mutliview Diffusion Model.
|
|
This&That: Language-Gesture Controlled Video Generation for Robot Planning
Boyang Wang,
Nikhil Sridhar,
Chao Feng,
Mark Van der Merwe,
Adam Fishman,
Nima Fazeli,
Jeong Joon Park
ICRA 2025
Paper
Ā· Website
Ā· Github
Ā· HuggingFace
First paper with Language & Gesture conditioned Video Generative Model for the robot action mentoring.
|
|
APISR: Anime Production Inspired Real-World Anime Super-Resolution
Boyang Wang,
Fengyu Yang,
Xihang Yu,
Chao Zhang,
Hanbin Zhao
CVPR 2024
Paper
Ā· Github
Ā· HuggingFace
SOTA Anime Super-Resolution and Restoration.
|
|
VCISR: Blind Single Image Super-Resolution with Video Compression Synthetic Data
Boyang Wang*,
Bowen Liu*,
Shiyu Liu*,
Fengyu Yang
WACV 2024
Paper
Ā· Github
First paper studying the Video Compression Degradation on Image Super-Resolution.
|
Teaching Assistant Experience
|
Something want to share at the end
|
I believe that the optimizer to the life is the faith one has.
|
|