论文标题
FOF:学习单眼实时重建的傅里叶占用领域
FOF: Learning Fourier Occupancy Field for Monocular Real-time Human Reconstruction
论文作者
论文摘要
深度学习的出现导致了人类重建的重大进展。但是,现有表示形式,例如参数模型,体素电网,网格和隐式神经表示,难以同时实现高质量的结果和实时速度。在本文中,我们提出了一种新型的强大,高效和灵活的3D表示,用于实时和准确的人类重建。 FOF代表一个3D对象,其与视图方向有2D字段正交,在每个2D位置,沿视图方向的对象的占用场与傅立叶级数的前几个术语紧凑,该序列的前几个术语将保留在2D域中的拓扑和邻域关系。 FOF可以存储为多通道图像,该图像与2D卷积神经网络兼容,可以弥合3D几何图像和2D图像之间的间隙。 FOF非常灵活且可扩展,例如,可以轻松地将参数模型集成到FOF中,以作为生成更健壮的结果。基于FOF,我们设计了前30+FPS高保真实时单眼重建框架。我们在公共数据集和实际捕获的数据上都证明了FOF的潜力。该代码将出于研究目的发布。
The advent of deep learning has led to significant progress in monocular human reconstruction. However, existing representations, such as parametric models, voxel grids, meshes and implicit neural representations, have difficulties achieving high-quality results and real-time speed at the same time. In this paper, we propose Fourier Occupancy Field (FOF), a novel powerful, efficient and flexible 3D representation, for monocular real-time and accurate human reconstruction. The FOF represents a 3D object with a 2D field orthogonal to the view direction where at each 2D position the occupancy field of the object along the view direction is compactly represented with the first few terms of Fourier series, which retains the topology and neighborhood relation in the 2D domain. A FOF can be stored as a multi-channel image, which is compatible with 2D convolutional neural networks and can bridge the gap between 3D geometries and 2D images. The FOF is very flexible and extensible, e.g., parametric models can be easily integrated into a FOF as a prior to generate more robust results. Based on FOF, we design the first 30+FPS high-fidelity real-time monocular human reconstruction framework. We demonstrate the potential of FOF on both public dataset and real captured data. The code will be released for research purposes.