Publications
You can also find a list of my publications on Google Scholar.
Published in CVPR, 2024
This paper presents a novel neural rendering based method for object inverse rendering by utilizing appearance under near- and far-field light sources.
Recommended citation: Fan Fei, Jiajun Tang, Ping Tan, and Boxin Shi. "VMINer: Versatile Multi-view Inverse Rendering with Near- and Far-field Light Sources." CVPR , 2024.
Published in NeurIPS, 2023
This paper presents a novel illumination-aware conditional image repainting method for lighting-realistic image generation.
Recommended citation: Jiajun Tang, Haofeng Zhong, Shuchen Weng, and Boxin Shi. "LuminAIRe: Illumination-Aware Conditional Image Repainting for Lighting-Realistic Generation." NeurIPS , 2023.
Published in CVPR, 2023
This paper presents a novel outdoor scene relighting method combining advantages of both NeRF-based methods and CNN-based methods.
Recommended citation: Siqi Yang, Xuanning Cui, Yongjie Zhu, Jiajun Tang, Si Li, Zhaofei Yu, and Boxin Shi. "Complementary Intrinsics from Neural Radiance Fields and CNNs for Outdoor Scene Relighting." CVPR, 2023.
Published in ECCV, 2022
This paper presents a novel disentangled and editable lighting representation for spatially-varying outdoor lighting with few parameters, which can be used for outdoor spatially-varying lighting estimation task.
Recommended citation: Jiajun Tang, Yongjie Zhu, Haoyu Wang, Jun Hoong Chan, Si Li, and Boxin Shi. "Estimating Spatially-Varying Lighting in Urban Scenes with Disentangled Representation." ECCV, 2022.
Published in AAAI, 2022
This paper propose a language-guided grey image colorization method which handles the color-object coupling and mismatch issues.
Recommended citation: Shuchen Weng, Hao Wu, Zheng Chang, Jiajun Tang, Si Li, and Boxin Shi. "L-CoDe: Language-based Colorization using Color-object Decoupled Conditions." AAAI, 2022.
Published in ICCP, 2021
This paper propose a method to decompose the shadow-free albedo and latent lighting, and render shape-(in)dependent shadings, given a single image of an outdoor urban scene, trained in a self-supervised manner.
Recommended citation: Yongjie Zhu, Jiajun Tang, Si Li, and Boxin Shi. "DeRenderNet: Intrinsic Image Decomposition of Urban Scenes with Shape-(In)dependent Shading Rendering." ICCP, 2021.
Published in ICCV, 2019
This paper improves the diversity and descriptiveness of machine-generated image captions by utilizing visual paragraphs in existing captioning datasets.
Recommended citation: Lixin Liu, Jiajun Tang, Xiaojun Wan, and Zongming Guo. "Generating Diverse and Descriptive Image Captions Using Visual Paraphrases." ICCV, 2019.
Last updated: Apr. 26, 2024