Skip to main content

TextureNet: Consistent Local Parametrizations for Learning From High-Resolution Signals on Meshes

Author(s): Huang, Jingwei; Zhang, Haotian; Yi, Li; Funkhouser, Thomas; Niebner, Matthias; et al

Download
To refer to this page use: http://arks.princeton.edu/ark:/88435/pr1fk2g
Abstract: We introduce, TextureNet, a neural network architecture designed to extract features from high-resolution signals associated with 3D surface meshes (e.g., color texture maps). The key idea is to utilize a 4-rotational symmetric(4-RoSy) field to define a domain for convolution on a surface. Though 4-RoSy fields have several properties favor-able for convolution on surfaces (low distortion, few singularities, consistent parameterization, etc.), orientations are ambiguous up to 4-fold rotation at any sample point. So, we introduce a new convolutional operator invariant to the4-RoSy ambiguity and use it in a network to extract features from high-resolution signals on geodesic neighborhoods of a surface. In comparison to alternatives, such as PointNet-based methods which lack a notion of orientation, the coherent structure given by these neighborhoods results in significantly stronger features. As an example application, we demonstrate the benefits of our architecture for 3D semantic segmentation of textured 3D meshes. The results show that our method outperforms all existing methods on the basis of mean IoU by a significant margin in both geometry-only(6.4%) and RGB+Geometry (6.9-8.2%) settings.
Publication Date: 2019
Citation: Huang, Jingwei, Haotian Zhang, Li Yi, Thomas Funkhouser, Matthias Nießner, and Leonidas J. Guibas. "TextureNet: Consistent Local Parametrizations for Learning From High-Resolution Signals on Meshes." In IEEE/CVF Conference on Computer Vision and Pattern Recognition (2019): pp. 4435-4444. doi:10.1109/CVPR.2019.00457
DOI: 10.1109/CVPR.2019.00457
ISSN: 1063-6919
EISSN: 2575-7075
Pages: 4435 - 4444
Type of Material: Conference Article
Journal/Proceeding Title: IEEE/CVF Conference on Computer Vision and Pattern Recognition
Version: Author's manuscript



Items in OAR@Princeton are protected by copyright, with all rights reserved, unless otherwise indicated.