Yeongtak Oh
Hi, I'm a fourth-year Ph.D. candidate in ECE at Seoul National University, working in the DSAIL Lab. I research computer vision and multi-modal reasoning. My work primarily explores post-training of generative models. I'm deeply interested in advancing multi-modal AI systems in more expressive and personalized ways.
I received my B.S. (2018) and M.S. (2020) degrees in Mechanical Engineering from Seoul National University. In 2021, I served as a Military Science and Technology Researcher at the AI R&D Center of the Korea Military Academy.
Email  / 
CV  / 
Google Scholar  / 
Github
|
|
News
[2025/09] One paper got accepted to NeurIPS 2025 Conference!
[2024/11] One paper got accepted to IJCV 2024 Journal!
[2024/07] One paper got accepted to BMVC 2024 Conference!
[2024/07] One paper got accepted to ECCV 2024 Conference
Conferences
|
RePIC: Reinforced Post-Training for Personalizing Multi-Modal Language Models
Yeongtak Oh, Jisoo Mok, Dohyun Chung, Juhyeon Shin, Sangha Park, Johan Barthelemy, and Sungroh Yoon
Neural Information Processing Systems (NeurIPS), 2025
project page /
arXiv
We propose RePIC, a reinforced post-training framework that outperforms SFT-based methods in multi-concept personalized image captioning by enhancing visual recognition and generalization through reward templates and curated instructions.
|
|
ControlDreamer: Stylized 3D Generation
with Multi-View ControlNet
Yeongtak Oh*, Jooyoung Choi*, Yongsung Kim, Minjun Park, Chaehun Shin, and Sungroh Yoon
* Equal Contribution
British Machine Vision Conference (BMVC), 2024
project page /
arXiv
ControlDreamer enables high-quality 3D generation with creative geometry and styles via multi-view ControlNet.
|
Journals
Preprints
|
Style-Friendly SNR Sampler for Style-Driven Generation
Jooyoung Choi*, Chaehun Shin*, Yeongtak Oh, Heeseung Kim, and Sungroh Yoon
* Equal Contribution
arxiv, 2024
project page /
arXiv
We propose the Style-friendly SNR sampler, which aggressively shifts the signal-to-noise ratio (SNR) distribution toward higher noise levels during fine-tuning to focus on noise levels where stylistic features emerge.
|
Talks
[2023.08.25] Recent Trends of Generative models in 3D vision
[2024.11.27] Image-Inversion of Diffusion Models
|