Higher-Order Function Networks for Learning Composable 3D Object RepresentationsDownload PDF

Published: 20 Dec 2019, Last Modified: 23 Mar 2025ICLR 2020 Conference Blind SubmissionReaders: Everyone
TL;DR: Neural nets can encode complex 3D objects into the parameters of other (surprisingly small) neural nets
Abstract: We present a new approach to 3D object representation where a neural network encodes the geometry of an object directly into the weights and biases of a second 'mapping' network. This mapping network can be used to reconstruct an object by applying its encoded transformation to points randomly sampled from a simple geometric space, such as the unit sphere. We study the effectiveness of our method through various experiments on subsets of the ShapeNet dataset. We find that the proposed approach can reconstruct encoded objects with accuracy equal to or exceeding state-of-the-art methods with orders of magnitude fewer parameters. Our smallest mapping network has only about 7000 parameters and shows reconstruction quality on par with state-of-the-art object decoder architectures with millions of parameters. Further experiments on feature mixing through the composition of learned functions show that the encoding captures a meaningful subspace of objects.
Keywords: computer vision, 3d reconstruction, deep learning, representation learning
Data: [ShapeNet](https://paperswithcode.com/dataset/shapenet)
Community Implementations: [![CatalyzeX](/images/catalyzex_icon.svg) 1 code implementation](https://www.catalyzex.com/paper/higher-order-function-networks-for-learning/code)
Original Pdf: pdf
8 Replies

Loading

OpenReview is a long-term project to advance science through improved peer review with legal nonprofit status. We gratefully acknowledge the support of the OpenReview Sponsors. © 2025 OpenReview