Yufan Zhou
Research Scientist at Adobe Research
Email: yufanzho AT buffalo DOT eduBio
Currently my research focuses on generative models. More specifically, I'm interested in:
- Multi-modal generative models (assistants) which are more user-friendly;
- Customizing pre-trained generative models;
- Saving the training or dataset construction cost in generative modeling;
I obtained my Ph.D. from the Department of Computer Science and Engineering, University at Buffalo, under the supervision of
Prof. Jinhui Xu and
Prof. Changyou Chen.
I received my B.E. degree from Zhejiang University.
I worked as a Research Intern with Chunyuan Li (Microsoft), Ruiyi Zhang (Adobe), Bingchen Liu (ByteDance).
Students who are interested in interning at Adobe, or seeking research collaborations, feel free to reach out to me.
News
- One paper accepted by AAAI 2025.
- One paper accepted by WACV 2025.
- One paper accepted by EMNLP 2024.
- One paper accepted by CoLM 2024.
- One paper accepted by ACL 2024.
- Two papers accepted by CVPR 2024.
- One paper accepted by ICLR 2024.
Selected Papers [More]
-
A zero-shot method for video customization, which can generate creative videos for user-input subject image, with desired style, color, texture, background required by user-input text.
-
Efficient method to construct dataset for subject-driven T2I generation, which can save at least tens of thousands of GPU hours.
-
IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2024.An assistant which can generate creative images for specific user-input subject along with text explanation and elaboration in 2-5 seconds, without any fine-tuning.
-
A novel framework for customized text-to-image generation without the use of regularization.
We can efficiently customize a large-scale text-to-image generation model on single GPU, with only one image provided by the user. -
IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2023.We propose a method termed Corgi, which can better generate image embeddings from text inside multimodal embedding space.
It benefits both standard and language-free text-to-image generation. And yes, I do have a Corgi. -
IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2022.Our proposed work, Lafite, is the first work which can successfully train text-to-image generation model with image-only dataset.
-
AAAI conference on Artificial Intelligence (AAAI), 2022.
-
International Conference on Learning Representations (ICLR), 2021.
-
Conference on Neural Information Processing Systems (NeurIPS), 2020.
Professional Service
- Conferences Program Committee/Reviewer: NeurIPS; ICML; ICLR; CVPR; AISTATS; AAAI; IJCAI; EMNLP; ECCV; ACL;
- Journal Reviewer: IEEE Transactions on Pattern Analysis and Machine Intelligence; IEEE Transactions on Neural Networks and Learning Systems; IEEE Transactions on Circuits and Systems for Video Technology;