Yihao LIU
| Google scholar | Github | XiaoHongShu |
I am a Research Scientist at the Shanghai Artificial Intelligence Laboratory, where I lead a team focusing on multimodal generation and understanding. I earned my Bachelor’s degree in 2018 and my Ph.D. in 2023, both from the University of Chinese Academy of Sciences (UCAS). During my doctoral studies, I was affiliated with the Shenzhen Institutes of Advanced Technology (SIAT), Chinese Academy of Sciences, under the supervision of Prof. Yu Qiao and Prof. Chao Dong. My research lies at the intersection of computer vision, generative modeling, and scientific intelligence, with particular emphasis on multimodal foundation models and image/video enhancement.
Throughout my student journey, I have been honored with prestigious awards, including the President’s Award of the Chinese Academy of Sciences, the Zhu Li Yue Hua Outstanding Doctoral Student Award, the CAS Excellent Youth League Member Award, the Beijing Outstanding Graduate Award, the SIAT President’s Innovation Award, as well as the CVMJ 2025 Best Paper Honorable Mention Award.
I have also excelled in multiple international and national competitions, such as 1st place in the PIRM 2018 Perceptual Image Super-Resolution Challenge, 1st place in the AIM 2020 Video Frame Interpolation Challenge, 2nd place in the NTIRE 2021 HDR Enhancement Challenge, 3rd place in the UDC 2020 Under-Display Camera Restoration Challenge. I serve as a reviewer for various top journals and conferences, including TPAMI, TIP, TCSVT, TMM, CVPR, ICCV, ECCV, NeurIPS, etc.
Current Research Focus
My current research focuses on pioneering a new generation of multimodal foundation models that integrate generation and understanding within a unified architecture. Specifically:
-
Unified Multimodal Architectures: Designing new-generation frameworks (e.g., discrete diffusion, autoregressive hybrids) that integrate text, image, video, and audio tasks, enabling coherent cross-modal representation, reasoning, and generation. -
Knowledge-Driven and Causality-Aware Modeling: Embedding structured world knowledge, physical realism, and causal reasoning into multimodal models, moving beyond perceptual fidelity toward scientifically grounded and logically consistent outputs. -
General Low-Level Vision Models: Consolidating diverse low-level vision tasks — restoration, enhancement, style transfer, and dense prediction — into a robust multimodal framework, advancing detail recovery, fidelity, and generalization for real-world applications. -
Post-training and Reward Alignment: Developing multimodal alignment and reinforcement learning paradigms, incorporating human preference modeling and expert feedback, to ensure outputs that are not only high-quality and aesthetic but also reliable, interpretable, and scientifically valid.
I am open to collaboration and discussions. Feel free to reach out at liuyihao@pjlab.org.cn or liuyihao14@mails.ucas.ac.cn
news
| Sep 10, 2025 | We are excited to announce Lumina-DiMOO, our latest unified multimodal generation and understanding model built upon an advanced discrete diffusion architecture. This framework demonstrates the strong potential of multimodal diffusion large language models (dLLM) to unify diverse tasks within a single, streamlined architecture, while delivering state-of-the-art performance that surpasses many existing unified models. Learn more and explore resources: [Homepage] [GitHub] [HuggingFace]. |
|---|---|
| Sep 01, 2025 | We introduce ArtiMuse, a multimodal large language model (MLLM) for professional aesthetic understanding, which is trained on ArtiMuse-10K, a meticulously curated, expert-annotated dataset. ArtiMuse-10K systematically defines eight explainable and fine-grained aesthetic attributes (e.g., Composition & Design, Visual Elements & Structure), with a wide coverage of diverse visual domains, including graphic design, 3D design, AIGC-generated images, photography, and painting & calligraphy. [Paper] [Homepage] [GitHub] [Online Demo v1.0] Note: ArtiMuse was officially released at WAIC 2025, in the forum “Evolving with AI: The Iteration and Resilience of Artistic Creativity”. |
| Jun 26, 2025 | Our video restoration method DiffVSR was accepted by ICCV2025. [Paper] [Homepage] |
| Apr 22, 2025 | Our video colorization method TCVC has won the CVMJ 2025 Best Paper Honorable Mention Award. |
| Apr 01, 2025 | We present Lunima-OmniLV (abbreviated as OmniLV), a universal multimodal multi-task framework for low-level vision that addresses over 100 sub-tasks across four major categories, including image restoration, image enhancement, weak-semantic dense prediction, and stylization. [Paper] [Homepage] |
| Jul 18, 2024 | GenLV was accepted by ACM MM2024. GenLV is a successive work of PromptGIP, which further broadens the tasks and improves performance. The paper can be found at here. |
| Jul 01, 2024 | Two papers were accepted by ECCV2024. By analyzing the relationships between image degradations, GRIDS propose a grouped learning method to deal with multiple-degradation restoration. X-Restormer is a new general image restoration backbone network, which possesses good task generality and achieves competitive performance across a variety of restoration tasks. |
| May 02, 2024 | PromptGIP was accepted by ICML2024. PromptGIP is a universal model for general image processing that covers image restoration, image enhancement, image feature extraction tasks, etc. Code is available at here. |
selected publications
- ECCV
GRIDS: Grouped Multiple-Degradation Restoration with Image Degradation SimilarityIn European Conference on Computer Vision, 2024 - ECCV
A Comparative Study of Image Restoration Networks for General Backbone Network DesignIn European Conference on Computer Vision, 2024 - ICML
Unifying Image Processing as Visual Prompting Question AnsweringIn Proceedings of the 41st International Conference on Machine Learning (ICML), 2024 - ACM MM
Learning A Low-Level Vision Generalist via Visual Task PromptIn Proceedings of the 32nd ACM International Conference on Multimedia, 2024 - CVMJ
Temporally Consistent Video Colorization with Deep Feature Propagation and Self-Regularization LearningComputational Visual Media, 2024 - CVPR
DegAE: A New Pretraining Paradigm for Low-Level VisionIn Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023 - TPAMI
Evaluating the Generalization Ability of Super-Resolution NetworksIEEE Transactions on pattern analysis and machine intelligence, 2023 - CVPR
Masked Image Training for Generalizable Deep Image DenoisingIn Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023 - TPAMI
CP3: Unifying Point Cloud Completion by Pretrain-Prompt-Predict ParadigmIEEE Transactions on Pattern Analysis and Machine Intelligence, 2023 - TMM
Very Lightweight Photo Retouching Network with Conditional Sequential ModulationIEEE Transactions on Multimedia, 2022 - TPAMI
Blind Image Super-Resolution: A Survey and BeyondIEEE transactions on pattern analysis and machine intelligence, 2022 - TPAMI
RankSRGAN: Super Resolution Generative Adversarial Networks with Learning to RankIEEE Transactions on Pattern Analysis and Machine Intelligence, 2021 - TPAMI
Interactive Multi-Dimension Modulation for Image RestorationIEEE Transactions on Pattern Analysis and Machine Intelligence, 2021 - ICCV
Learn to Match: Automatic Matching Network Design for Visual TrackingIn International Conference on Computer Vision (ICCV), 2021 - arXiv
Discovering" Semantics" in Super-Resolution NetworksarXiv preprint arXiv:2108.00406, 2021 - AAAI
FD-GAN: Generative Adversarial Networks with Fusion-Discriminator for Single Image DehazingIn Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), 2020 - ECCV
Conditional Sequential Modulation for Efficient Global Image RetouchingIn European Conference on Computer Vision (ECCV), 2020 - ECCVW
Enhanced Quadratic Video InterpolationIn European Conference on Computer Vision (ECCV) Workshops, 2020 - ICCVRankSRGAN: Generative Adversarial Networks with Ranker for Image Super-ResolutionIn International Conference on Computer Vision (ICCV), 2019
- ECCVWESRGAN: Enhanced Super-Resolution Generative Adversarial NetworksIn Proceedings of the European conference on computer vision (ECCV) workshops, 2018