We propose CorrespondentDream, an effective method to leverage annotation-free, cross-view correspondences yielded from the diffusion U-Net to provide additional 3D prior to the NeRF optimization process. We find that these correspondences are strongly consistent with human perception, and by adopting it in our loss design, we are able to produce NeRF models with geometries that are more coherent with common sense, e.g., more smoothed object surface, yielding higher 3D fidelity. We demonstrate the efficacy of our approach through various comparative qualitative results.
CorrespondentDream generates multi-view consistent objects and scenes with improved 3D fidelity
Zoom in for clearer visualization of the infidelities in 3D geometry such as concavities / missing surfaces.