Boyang Wang

Hello! This is Boyang Wang (ēŽ‹åšę“‹). I am an incoming PhD student to University of Virginia supervised by Zezhou Cheng. I was graduated with M.S. of Robotics and B.S. of Computer Science from University of Michigan (UMich). At UMich, I worked with Prof. JJ Park, Prof. Nima Fazeli, Prof. Hun-Seok Kim, and Prof. Chad Jenkins.

My research interests mainly lie in Generative Models (especially video) for General, Robotics, and Anime domain!

Email  Ā·  Twitter  Ā·  Github  Ā·  Google Scholar

profile photo
News

  • 05/2025: 🎉 Frame In-N-Out: Unbounded Controllable Image-to-Video Generation is released on Arxiv.
  • 02/2025: 🎉 One paper is accepted by CVPR 2025. See you in Nashville!
  • 01/2025: 🎉 One paper is accepted by ICRA 2025. See you in Atlanta!
  • 01/2025: 🎉 One paper is accepted by RA-L.
  • 12/2024: 🎉 Graduated from University of Michigan as Master Degree.
  • 08/2024: 🎉 Being selected as GSI for the EECS498 Computer Graphics and Generative Models.
  • 07/2024: 🎉 This&That: Language-Gesture Controlled Video Generation for Robot Planning is released on Arxiv.
  • 02/2024: 🎉 One paper is accepted by CVPR 2024.
  • 10/2023: 🎉 One paper is accepted by WACV 2024.

  • Selected Publications

    (* indicates co-author)



    Frame In-N-Out: Unbounded Controllable Image-to-Video Generation
    Boyang Wang, Xuweiyi Chen, Matheus Gadelha, Zezhou Cheng
    Under Review
    Paper Ā· Website

    First paper in Video Generation that studies Frame In and Frame Out cinematics effect, which is conditioned on First Frame Expanded Canvas, Text Prompt, Motion Trajectory and Identity Reference.

    Sparse Image Sets Restoration with Multi-View Diffusion Model
    Yucheng Mao *, Boyang Wang *, Nilesh Kulkarni, Jeong Joon Park
    CVPR 2025
    Paper Ā· Website

    Reconstruct a 3D scene with degraded image sets by Mutliview Diffusion Model.

    This&That: Language-Gesture Controlled Video Generation for Robot Planning
    Boyang Wang, Nikhil Sridhar, Chao Feng, Mark Van der Merwe, Adam Fishman, Nima Fazeli, Jeong Joon Park
    ICRA 2025
    Paper Ā· Website Ā· Github Ā· HuggingFace

    First paper with Language & Gesture conditioned Video Generative Model for the robot action mentoring.

    APISR: Anime Production Inspired Real-World Anime Super-Resolution
    Boyang Wang, Fengyu Yang, Xihang Yu, Chao Zhang, Hanbin Zhao
    CVPR 2024
    Paper Ā· Github Ā· HuggingFace

    SOTA Anime SR and Restoration.

    VCISR: Blind Single Image Super-Resolution with Video Compression Synthetic Data
    Boyang Wang*, Bowen Liu*, Shiyu Liu*, Fengyu Yang
    WACV 2024
    Paper Ā· Github

    First paper studying the Video Compression Degradation on Image SR.


    Teaching Assistant Experience

  • Graduate Student Instructor (GSI), EECS498: Computer Graphics and Generative Models, Fall 2024
  • Graduate Student Instructor (GSI), EECS367/ROB380/ROB511: AutoRob, Winter 2024
  • Teaching Assistant (IA), EECS367/ROB380/ROB511: AutoRob, Winter 2023

  • Intern Work Experience

  • Machine Learning Engineer, ByteDance , 04/2022 - 08/2022

  • Something want to share at the end

    I believe that the optimizer to the life is the faith one has.