I am an assistant professor, Ph.D. advisor at the School of Artificial Intelligence, Shanghai Jiao Tong University. I obtained my Ph.D. at the Department of Electronic and Computer Engineering, National University of Singapore in June 2025, supervised by Prof. Xinchao Wang. Before that, I received my bachelorβs degree in the Department of Computer Science and Technology from Nanjing University in 2021. I was also fortunate to work as a research intern at Rutgers University under the supervision of Prof. Hao Wang and at Baidu Inc. under the supervision of Tianwei Lin.
My research interests include visual synthesis and training with synthetic data. Since 2020, I have published over 10 first-author research papers at top journals/conferences on these topics. Multiple works are selected as paper award nominations, oral, highlight, and spotlight. During my doctoral studies, I received the NUS Research Scholarship and the 2023 National Award for Outstanding Self-financed Chinese Students Abroad.
Currently, I am creating Learning and Synthesis Hub and recruiting Ph.D. students (2~3 every year starting from fall 2026), Master students (1~2 every year starting from fall 2026), and multiple research interns. If you are interested in working with me at Shanghai Jiao Tong University on the following topics:
feel free to reach out via liusonghua{AT}sjtu.edu.cn / liusonghua_site{AT}163.com. We are always looking for self-motivated students actively.
Currently, I have run out of Ph.D. positions for 2026.
Last Update: 2025.07.01
[2025.06] π₯We release WinWinLay, a training-free layout-to-image method that advances both worlds of precision and realism.
[2025.06] π₯3 papers are accepted by ICCV 2025! Congratulations to all co-authors!π
[2025.05] π₯We release IEAP: Your Free GPT-4o Image Editor. It decouples a complex image editing instruction into several programmable atomic operations and achieves compelling performance! Codes are available here!π Demo is available here.π€
[2025.05] π₯We release LLaVA-Meteor, a novel, efficient, and promising method to process visual tokens in VLM.
[2025.05] Congratulations to Mr. Xinhao Tan for having his work accepted at ACL 2025 during his undergraduate period!π
[2025.05] π₯2 papers are accepted by ICML 2025! Congratulations to all co-authors!π
[2025.04] π₯We release Flash Sculptor, an efficient and scalable solution to reconstruct a complex and compositional 3D scene from merely a single image! Try it hereπ
[2025.03] π₯We release URAE: Your Free FLUX Pro Ultra. It adapts pre-trained diffusion transformers like FLUX to 4K resolution with merely 3K training samples! Codes are available here!π Demos are available here and here.π€
[2025.03] π₯We release OminiControl2! OminiControl1 has achieved impressive results on controllable generation. Now, OminiControl2 makes it efficient as well!
[2025.03] π₯We release CFM! Spectral Filtering provides dataset distillation with a unified perspective!
[2025.02] π₯3 papers are accepted by CVPR 2025! Congratulations to all co-authors!π
[2025.02] π₯We release a benchmark for large-scale dataset compression here. Try it here!π
[2024.12] π₯We release CLEAR, a simple-yet-effective strategy to linearize the complexity of pre-trained diffusion transformers, such as FLUX and SD3. Try it here!π
[2024.12] π₯We release OminiControl, a minimal yet powerful universal control framework for Diffusion Transformer models like FLUX. Try it here!π