As a software engineer (ML R&D) on the XR team at Ohouse, I specialize in 3D vision and generative models. I played a key role in developing AR solutions such as the 'Space Saving' and 'Furniture Removal' features, which enable users to visualize their space with accurately scaled 3D models and experiment with virtual furniture placements. I worked closely with Hyowon Ha and Jaeheung surh to enhance user experiences in extended reality. Previously, I worked as a research scientist on the generation research team at NAVER AI Lab, collaborating with Junho Kim, Gayoung Lee, and Yunjey Choi, as well as Prof. Jun-Yan Zhu (CMU). I received my B.S. and M.S degrees in computer science and engineering from Seoul National University in 2019 and 2021, respectively. During my Master’s program, I was advised by Prof. Sungjoo Yoo. My product/research interests are generative models and 3D.
When purchasing furniture, do you wonder if it will match your space or if the size is right? Try out these new features!
Capture your space, and it will be converted into a 3D model with accurate dimensions. You can virtually place new furniture to see how it fits.
Remove existing furniture virtually without physically moving it and visualize your space with new furniture in place.
Most recent publications on Google Scholar.
‡ indicates equal contribution.
3D-aware Blending with Generative NeRFs
Hyunsu Kim, Gayoung Lee, Yunjey Choi, Jin-Hwa Kim, Jun-Yan Zhu
ICCV, 2023
BallGAN: 3D-aware Image Synthesis with a Spherical Background
Minjung Shin, Yunji Seo, Jeongmin Bae, Young Sun Choi, Hyunsu Kim, Hyeran Byun, Youngjung Uh
ICCV, 2023
Diffusion Models with Grouped Latents for Interpretable Latent Space
Sangyun Lee, Gayoung Lee, Hyunsu Kim, Junho Kim, Youngjung Uh
ICMLW, 2023
User-friendly Image Editing with Minimal Text Input: Leveraging Captioning and Injection Techniques
Sunwoo Kim, Wooseok Jang, Hyunsu Kim, Junho Kim, Yunjey Choi, Seungryong Kim, Gayeong Lee
arXiv, 2023
Diffusion Video Autoencoders: Toward Temporally Consistent Face Video Editing via Disentangled Video Encoding
Gyeongman Kim, Hajin Shim, Hyunsu Kim, Yunjey Choi, Junho Kim, Eunho Yang
CVPR, 2023
Context-Preserving Two-Stage Video Domain Translation for Portrait Stylization
Doyeon Kim, Eunji Ko, Hyunsu Kim, Yunji Kim, Junho Kim, Dongchan Min, Junmo Kim, Sung Ju Hwang
CVPRW, 2023
Learning Input-agnostic Manipulation Directions in StyleGAN with Text Guidance
Yoonjeon Kim, Hyunsu Kim, Junho Kim, Yunjey Choi, Eunho Yang
ICLR 2023
Generator Knows What Discriminator Should Learn in Unconditional GANs
Gayoung Lee, Hyunsu Kim, Junho Kim, Seonghyeon Kim, Jung-Woo Ha, Yunjey Choi
ECCV 2022
Generating videos with dynamics-aware implicit generative adversarial networks
Sihyun Yu‡, Jihoon Tack‡, Sangwoo Mo‡, Hyunsu Kim, Junho Kim, Jung-Woo Ha, Jinwoo Shin
ICLR 2022
StyleMapGAN: Exploiting Spatial Dimensions of Latent in GAN for Real-time Image Editing
Hyunsu Kim, Yunjey Choi, Junho Kim, Sungjoo Yoo, Youngjung Uh
CVPR 2021
Tag2pix: Line art colorization using text tag with secat and changing loss
Hyunsu Kim‡, Ho Young Jhoo‡, Eunhyeok Park, Sungjoo Yoo
ICCV 2019