Enhancing 2D Representation Learning with a 3D Prior

Type: Preprint

Publication Date: 2024-06-04

Citations: 0

DOI: https://doi.org/10.48550/arxiv.2406.02535

Abstract

Learning robust and effective representations of visual data is a fundamental task in computer vision. Traditionally, this is achieved by training models with labeled data which can be expensive to obtain. Self-supervised learning attempts to circumvent the requirement for labeled data by learning representations from raw unlabeled visual data alone. However, unlike humans who obtain rich 3D information from their binocular vision and through motion, the majority of current self-supervised methods are tasked with learning from monocular 2D image collections. This is noteworthy as it has been demonstrated that shape-centric visual processing is more robust compared to texture-biased automated methods. Inspired by this, we propose a new approach for strengthening existing self-supervised methods by explicitly enforcing a strong 3D structural prior directly into the model during training. Through experiments, across a range of datasets, we demonstrate that our 3D aware representations are more robust compared to conventional self-supervised baselines.

Locations

  • arXiv (Cornell University) - View - PDF

Similar Works

Action Title Year Authors
+ A Closer Look at Invariances in Self-supervised Pre-training for 3D Vision 2022 Lanxiao Li
Michael Heizmann
+ ShapeCodes: Self-Supervised Feature Learning by Lifting Views to Viewgrids 2017 Dinesh Jayaraman
Ruohan Gao
Kristen Grauman
+ CroCo: Self-Supervised Pre-training for 3D Vision Tasks by Cross-View Completion 2022 Philippe Weinzaepfel
Vincent Leroy
Thomas G. Lucas
Romain Brégier
Yohann Cabon
Vaibhav Arora
Leonid Antsfeld
Boris Chidlovskii
Gabriela Csurka
Jérôme Revaud
+ ShapeCodes: Self-Supervised Feature Learning by Lifting Views to Viewgrids 2017 Dinesh Jayaraman
Ruohan Gao
Kristen Grauman
+ PDF Chat MaskLRF: Self-supervised Pretraining via Masked Autoencoding of Local Reference Frames for Rotation-invariant 3D Point Set Analysis 2024 Takahiko Furuya
+ PDF Chat Pri3D: Can 3D Priors Help 2D Representation Learning? 2021 Ji Hou
Saining Xie
Benjamin Graham
Angela Dai
Matthias Niesner
+ Scaling and Benchmarking Self-Supervised Visual Representation Learning. 2019 Priya Goyal
Dhruv Mahajan
Abhinav Gupta
Ishan Misra
+ PDF Chat Scaling and Benchmarking Self-Supervised Visual Representation Learning 2019 Priya Goyal
Dhruv Mahajan
Abhinav Gupta
Ishan Misra
+ Scaling and Benchmarking Self-Supervised Visual Representation Learning 2019 Priya Goyal
Dhruv Mahajan
Abhinav Gupta
Ishan Misra
+ PDF Chat MaskLRF: Self-supervised Pretraining via Masked Autoencoding of Local Reference Frames for Rotation-invariant 3D Point Set Analysis 2024 Takahiko Furuya
+ Fostering Generalization in Single-view 3D Reconstruction by Learning a Hierarchy of Local and Global Shape Priors 2021 Jan Bechtold
Maxim Tatarchenko
Volker Fischer
Thomas Brox
+ Fostering Generalization in Single-view 3D Reconstruction by Learning a Hierarchy of Local and Global Shape Priors 2021 Jan Bechtold
Maxim Tatarchenko
Volker Fischer
Thomas Brox
+ PDF Chat Fostering Generalization in Single-view 3D Reconstruction by Learning a Hierarchy of Local and Global Shape Priors 2021 Jan Bechtold
Maxim Tatarchenko
Volker Fischer
Thomas Brox
+ PDF Chat Enhancing Generalizability of Representation Learning for Data-Efficient 3D Scene Understanding 2024 Yunsong Wang
Na Zhao
Gim Hee Lee
+ PDF Chat Probing the 3D Awareness of Visual Foundation Models 2024 Mohamed El Banani
Amit Raj
Kevis-Kokitsi Maninis
Abhishek Kar
Yuanzhen Li
Michael Rubinstein
Deqing Sun
Leonidas Guibas
Justin Johnson
Varun Jampani
+ PDF Chat Enhancing Generalizability of Representation Learning for Data-Efficient 3D Scene Understanding 2024 Yunsong Wang
Na Zhao
Gim Hee Lee
+ PDF Chat Multi-Task Self-Training for Learning General Representations 2021 Golnaz Ghiasi
Barret Zoph
Ekin D. Cubuk
Quoc V. Le
Tsung-Yi Lin
+ Multi-Task Self-Training for Learning General Representations 2021 Golnaz Ghiasi
Barret Zoph
Ekin D. Cubuk
Quoc V. Le
Tsung-Yi Lin
+ PDF Chat Multi-Task Self-Training for Learning General Representations 2021 Golnaz Ghiasi
Barret Zoph
Ekin D. Cubuk
Quoc V. Le
Tsung-Yi Lin
+ Pri3D: Can 3D Priors Help 2D Representation Learning? 2021 Ji Hou
Saining Xie
Benjamin Graham
Angela Dai
Matthias Nießner

Works That Cite This (0)

Action Title Year Authors

Works Cited by This (0)

Action Title Year Authors