We present RGB-D-Fusion, a multi-modal conditional denoising diffusion probabilistic model to generate high resolution depth maps from low-resolution monocular RGB images of humanoid subjects. RGB-D-Fusion first generates a low-resolution depth map using an image conditioned denoising diffusion probabilistic model and then upsamples the depth map using a second denoising diffusion probabilistic model conditioned on a low-resolution RGB-D image. We further introduce a novel augmentation technique, depth noise augmentation, to increase the robustness of our super-resolution model
Neural scene representations, such as Neural Radiance Fields (NeRF), are based on training a multila...
International audienceThis study explores the use of photometric techniques (shape-from-shading and ...
Estimating depth from RGB images is a long-standing ill-posed problem, which has been explored for d...
We propose DiffuStereo, a novel system using only sparse cameras (8 in this work) for high-quality 3...
We address the problem of people detection in RGB-D data where we leverage depth information to deve...
Performing super-resolution of a depth image using the guidance from an RGB image is a problem that ...
Monocular depth estimation is a challenging task that predicts the pixel-wise depth from a single 2D...
RGB-D data has turned out to be a very useful representation for solving fundamental computer visio...
Thesis (Ph.D.)--University of Washington, 2018With the introduction of economical depth cameras, com...
International audienceThis paper addresses the problem of range-stereo fusion, for the construction ...
Abstract—The emergence of low cost sensors capable of providing texture and depth information of a s...
Recently, diffusion model have demonstrated impressive image generation performances, and have been ...
International audienceInput HR RGB images and LR depth maps Output HR albedo and depth maps Relighti...
Neural scene representations, such as Neural Radiance Fields (NeRF), are based on training a multila...
University of Technology Sydney. Faculty of Engineering and Information Technology.With the developi...
Neural scene representations, such as Neural Radiance Fields (NeRF), are based on training a multila...
International audienceThis study explores the use of photometric techniques (shape-from-shading and ...
Estimating depth from RGB images is a long-standing ill-posed problem, which has been explored for d...
We propose DiffuStereo, a novel system using only sparse cameras (8 in this work) for high-quality 3...
We address the problem of people detection in RGB-D data where we leverage depth information to deve...
Performing super-resolution of a depth image using the guidance from an RGB image is a problem that ...
Monocular depth estimation is a challenging task that predicts the pixel-wise depth from a single 2D...
RGB-D data has turned out to be a very useful representation for solving fundamental computer visio...
Thesis (Ph.D.)--University of Washington, 2018With the introduction of economical depth cameras, com...
International audienceThis paper addresses the problem of range-stereo fusion, for the construction ...
Abstract—The emergence of low cost sensors capable of providing texture and depth information of a s...
Recently, diffusion model have demonstrated impressive image generation performances, and have been ...
International audienceInput HR RGB images and LR depth maps Output HR albedo and depth maps Relighti...
Neural scene representations, such as Neural Radiance Fields (NeRF), are based on training a multila...
University of Technology Sydney. Faculty of Engineering and Information Technology.With the developi...
Neural scene representations, such as Neural Radiance Fields (NeRF), are based on training a multila...
International audienceThis study explores the use of photometric techniques (shape-from-shading and ...
Estimating depth from RGB images is a long-standing ill-posed problem, which has been explored for d...