Offre de stage Master 2, équipe SIAM : « Leveraging Diffusion Models for Relightable and Editable Scene Representations »

/, Equipe SIAM, Offre Doctorat/Post-Doctorat/Stage, Plate-forme EVR@, Plate-formes, Recherche, Recrutement/Offre de stage Master 2, équipe SIAM : « Leveraging Diffusion Models for Relightable and Editable Scene Representations »

Offre de stage Master 2, équipe SIAM : « Leveraging Diffusion Models for Relightable and Editable Scene Representations »

Intership title : Leveraging Diffusion Models for Relightable and Editable Scene Representations

Project description

Recovering the properties of the scene such the shape, material and lighting properties is a crucial task in many computer vision and computer graphics applications. This task is referred to as inverse rendering, and it enables object insertion [1], scene relighting [2] and scene editing [3].

Recent advances in learning a 3D representation of the scene have shown impressive novel synthesis results such as NeRF [4] and 3D Gaussian Splatting [5]. However, these representations are not relightable /editable as the scene properties are baked into the radiance field. Many state-of-the-art solutions [6,7,8] have proposed inverse rendering pipelines making those 3D representations editable. Despite this progress, current methods often struggle with cast shadows, specular highlights, and other complex lighting interactions.

Diffusion-based generative models [9] have emerged as a promising visual generation approach. Diffusion models can change many images aspects such as the image style [10], or blending foreground object into the background [11], relighting scenes [12], editing the colors of specific objects [13] etc. This adaptability highlights diffusion models’ potential to learn and manipulate various intrinsic scene attributes effectively, including materials and lighting conditions, while maintaining photorealism. Their use in editing tasks results in representations that are flexible, allowing for manipulation of the scene properties [14, 15]. This internship will focus on developing methods that leverage diffusion models to disentangle and manipulate intrinsic scene properties, including materials, and lighting. The intern will explore novel approaches to produce representations that are fully editable and relightable. The specific objectives include:

  • Intrinsic scene decomposition: Investigate current state-of-the-art diffusion models that decompose images into core components, enabling photorealistic editing and relighting while preserving scene coherence.
  • Handling complex lighting interactions: Address current SOTA challenges in rendering complex lighting effects, such as realistic shadow and specular highlights.
  • Relightable and editable representation: develop novel framework that leverage diffusion models to produce high-quality 3D scene representations. These representations should allow for intrinsic scene properties, such as material and lighting, to be manipulated. The goal is to create models that are not only relightable and editable but also maintain high fidelity and realism in their outputs. This work will build on recent advancements in inverse rendering and diffusion model.
  • Evaluation: Evaluating the proposed techniques against existing state-of-the-art solutions, aiming to outperform them in terms of realism, fidelity, and editing across different lighting conditions.

 

Description of research activities:

  • Study state-of-the-art inverse rendering and relighting methods.
  • Identify bottlenecks in the current approaches of the state-of-the-art.
  • Propose new solutions to improve scene understanding and editing.
  • Research and develop algorithms based on the proposed solutions.
  • Apply proposed algorithms to existing datasets.
  • Publish research results at top conferences and participate in scientific seminars.

 

Supervision:

This internship will be supervised by as a PhD candidate working on NeRF-based inverse rendering and a Professor at University of Evry.

Prerequisites:              

The candidate should be motivated to carry out world-class research and should have a Master in Computer Vision and/or Robotics. They should have solid skills in the following domains:

  • Implement Code in Python, and CUDA (optional) using deep learning libraries (PyTorch, PyTorchlightning)
  • Good knowledge of Git and Linux systems
  • Project reporting and planning
  • Write scientific publications and participate in conferences
  • Fluency in spoken and written English and/or French is a plus
  • Intercultural and coordination skills, hands-on and can-do attitude Interpersonal skills, team spirit, and independent working style

 

Contacts and Application:

Hala DJEGHIM –

Désiré SIDIBE – drodesire.sidibe@univ-evry.f

 

Internship information: start in February/March 2024 for 6 months, located in University of Evry.

Application deadline:  January 10th, 2025.

Application Files: CV + motivation letter + transcript of records for academic years 2022-2023 and 20232024 + any other relevant document. To submit your application, send all documents in one PDF file at the contact addresses.

References:

[1] Wang, Z., et al., « Neural Fields Meet Explicit Geometric Representations for Inverse Rendering of Urban Scenes« ,  Conference on Computer Vision and Pattern Recognition (CVPR), 2023.

[2] Zhu, X., et al., « Relighting Scenes with Object Insertions in Neural Radiance Fields« , arXiv preprint arXiv: 14806, 2024.

[3] Zhu, J., et al.,  » I2-SDF: Intrinsic Indoor Scene Reconstruction and Editing via Raytracing in Neural SDFs« , CVPR, 2023

[4] Mildenhall, B., et al., « NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis, » ECCV, 2020.

[5] Kerbl, B., et al., « 3D Gaussian Splatting for Real-Time Radiance Field Rendering« , SIGGRAPH 2023.

[6] Zhanh, K., et al., « PhySG: Inverse Rendering with Spherical Gaussians for Physics-based Material Editing and Relighting » CVPR 2021.

[7] Zhang, X., et al., « NeRFactor: Neural Factorization of Shape and Reflectance under an Unknown Illumination« , SIGGRAPH Asia 2021.

[8] Liang, R., et al.,  » ENVIDR: Implicit Differentiable Renderer with Neural Environment Lighting« , ICCV, 2023.

[9] Ho, J., et al., « Denoising Diffusion Probabilistic Models » NeurIPS, 2020.

[10] Haque, A., et al.,  » Instruct-NeRF2NeRF« , ICCV

[11] Nichol, A., et al., « GLIDE: Towards Photorealistic Image Generation and Editing with Text-Guided Diffusion Models« , ICML, 2022.

[12] Kocsis, P ., et al.,  » LightIt: Illumination Modeling and Control for Diffusion Models« , CVPR, 2024.

[13] Liang, Z., et al.,  » Control Color: Multimodal Diffusion-Based Interactive Image Colorization« , arXiv preprint arXiv: 10855, 2024.

[14] Chen, X., et al., « IntrinsicAnything: Learning to Decompose and Edit Scenes from Images« , arxiv: 2404.11593, 2024.

[15] Jin, H., et al., « Neural Gaffer: Lighting and Material Editing via Neural Diffusion« , NeurIPS, 2024.

  • Date de l’appel : 13/11/2024
  • Statut de l’appel : Non pourvu
  • Contacts cotés IBISC : Désiré SIDIBE (PR Univ. Évry, IBISC équipe SIAM), Hala DJEGHIM (doctorante, IBISC équipe SIAM), desireDOTsidibeATuniv-evryDOTfr, halaDOTdjeghimATgmailDOTcom
  • Sujet de stage niveau Master 2 (format PDF)
  • Web équipe SIAM
WP to LinkedIn Auto Publish Powered By : XYZScripts.com