arXiv Analytics

Sign in

arXiv:1911.11938 [cs.CV]AbstractReferencesReviewsResources

Transfer Learning in Visual and Relational Reasoning

T. S. Jayram, Vincent Marois, Tomasz Kornuta, Vincent Albouy, Emre Sevgen, Ahmet S. Ozcan

Published 2019-11-27Version 1

Transfer learning is becoming the de facto solution for vision and text encoders in the front-end processing of machine learning solutions. Utilizing vast amounts of knowledge in pre-trained models and subsequent fine-tuning allows achieving better performance in domains where labeled data is limited. In this paper, we analyze the efficiency of transfer learning in visual reasoning by introducing a new model (SAMNet) and testing it on two datasets: COG and CLEVR. Our new model achieves state-of-the-art accuracy on COG and shows significantly better generalization capabilities compared to the baseline. We also formalize a taxonomy of transfer learning for visual reasoning around three axes: feature, temporal, and reasoning transfer. Based on extensive experimentation of transfer learning on each of the two datasets, we show the performance of the new model along each axis.

Related articles: Most relevant | Search more
arXiv:1608.08614 [cs.CV] (Published 2016-08-30)
What makes ImageNet good for transfer learning?
arXiv:1811.08737 [cs.CV] (Published 2018-11-21)
SpotTune: Transfer Learning through Adaptive Fine-tuning
arXiv:2007.04234 [cs.CV] (Published 2020-06-19)
Transfer Learning or Self-supervised Learning? A Tale of Two Pretraining Paradigms