22/11/2021

Self-Supervised Learning of Image Scale and Orientation

Jongmin Lee, Yoonwoo Jeong, Minsu Cho

Keywords: self-supervised learning, scale and orientation estimation, histogram alignment, patch pose estimation, visual correspondence, wide-baseline matching, multi-modal distribution

Abstract: We study the problem of learning to assign a characteristic pose, i.e., scale and orientation, for an image region of interest. Despite its apparent simplicity, the problem is non-trivial; it is hard to obtain a large-scale set of image regions with explicit pose annotations that a model directly learns from. To tackle the issue, we propose a self-supervised learning framework with a histogram alignment technique. We generate pairs of image regions by random scaling and rotating and then train a model to predict their distributions so that their relative difference is consistent with scaling and rotating being used. The proposed method learns a non-parametric and multi-modal distribution of scale and orientation without any supervision, achieving a significant improvement over previous methods in experimental evaluation.

 0
 0
 0
 0
This is an embedded video. Talk and the respective paper are published at BMVC 2021 virtual conference. If you are one of the authors of the paper and want to manage your upload, see the question "My papertalk has been externally embedded..." in the FAQ section.

Comments

Post Comment
no comments yet
code of conduct: tbd Characters remaining: 140

Similar Papers