高级检索
杨志豪, 彭伟龙, 唐可可, 汪洋涛, 谭恒良, 方美娥. 基于紧凑型线性混合半侧蒙皮权重的神经辐射场人体表示方法[J]. 计算机辅助设计与图形学学报. DOI: 10.3724/SP.J.1089.2024-00344
引用本文: 杨志豪, 彭伟龙, 唐可可, 汪洋涛, 谭恒良, 方美娥. 基于紧凑型线性混合半侧蒙皮权重的神经辐射场人体表示方法[J]. 计算机辅助设计与图形学学报. DOI: 10.3724/SP.J.1089.2024-00344
Yang Zhihao, Peng Weilong, Tang Keke, Wang Yangtao, Tan Hengliang, Fang Meie. Compact Half-Side Body Linear Blending Skinning Weights for Human NeRF Representation[J]. Journal of Computer-Aided Design & Computer Graphics. DOI: 10.3724/SP.J.1089.2024-00344
Citation: Yang Zhihao, Peng Weilong, Tang Keke, Wang Yangtao, Tan Hengliang, Fang Meie. Compact Half-Side Body Linear Blending Skinning Weights for Human NeRF Representation[J]. Journal of Computer-Aided Design & Computer Graphics. DOI: 10.3724/SP.J.1089.2024-00344

基于紧凑型线性混合半侧蒙皮权重的神经辐射场人体表示方法

Compact Half-Side Body Linear Blending Skinning Weights for Human NeRF Representation

  • 摘要: 基于单目视频序列的神经辐射场(NeRF)人体重建由于视野范围限制, 对动作和外观信息覆盖不足, 影响动作可编辑性和图像渲染质量. 为此, 提出了一种新颖的紧凑半侧身体权重的NeRF人体表征方法. 首先, 基于NeRF构建标准人体姿态模型; 然后, 引入可学习的半侧身体线性混合蒙皮权重, 利用人体的对称性提升动作的表达能力, 并实现从特定动作到标准姿态的精准转换; 最后, 通过结构化权重衰减和基于权重的采样策略, 提升体渲染效率和质量. 实验结果表明, 所提方法在ZJU-MoCap数据集上处理未见过的动作时表现出较强的可编辑性, 在SSIM, PSNR和LPIPS评价指标上分别为0.957, 29.16和0.028, 均优于对比方法.

     

    Abstract: Human body reconstruction using Neural Radiance Fields (NeRF) based on monocular video sequences faces challenges due to limited field of view, leading to insufficient coverage of motion and appearance information, which negatively impacts motion editability and image rendering quality. To address these issues, we propose a novel compact half-side weight human NeRF representation. First, a canonical human pose model is constructed based on NeRF; then, learnable half-side linear blend skinning weights are introduced to leverage human body symmetry to enhance motion expressiveness, enabling precise transformation from specific poses to canonical poses. Finally, a structured weight decay and weight-based sampling strategy is employed to improve volumetric rendering efficiency and quality. Experimental results show that the proposed method exhibits strong editability for unseen motions on the ZJU-MoCap dataset, achieving SSIM, PSNR, and LPIPS scores of 0.957, 29.16, and 0.028, respectively, outperforming the compared methods in all metrics.

     

/

返回文章
返回