Learning Deep Asymmetric Tolerant Part Representation
File version
Author(s)
Zhao, Y
Gao, Y
Xiong, S
Griffith University Author(s)
Primary Supervisor
Other Supervisors
Editor(s)
Date
Size
File type(s)
Location
License
Abstract
Categorization objects at a sub-ordinate level inevitably poses a significant challenge, i.e. , inter-class difference is very subtle and only exists in a few key parts. Therefore, how to localize these key parts for discriminative visual categorization without requiring expensive pixel-level annotations becomes a core question. To that end, this paper introduces a novel asymmetry tolerant part segmentation network (ATP-Net). ATP-Net simultaneously learns to segment parts and identify objects in an end-to-end manner using only image-level category labels. Given the intrinsic asymmetry property of part alignment, a desirable learning of part segmentation should be capable of incorporating such property. Despite the efforts towards regularizing weakly supervised part segmentation, none of them consider this vital and intrinsic property, i.e. , the spatial asymmetry of part alignment. Our work, for the first time, proposes to explicitly characterize the spatial asymmetry of part alignment for visual tasks. We propose a novel asymmetry loss function to guide the part segmentation by encoding the spatial asymmetry of part alignment, i.e. , restricting the upper bound of how asymmetric those self-similar parts are to each other in the network learning. Via a comprehensive ablation study, we verify the effectiveness of the proposed ATP-Net in driving the network learning towards semantically meaningful part segmentation and discriminative visual categorization. Consistently superior/competitive performance are reported on 12 datasets covering crop cultivar classification, plant disease classification, bird/butterfly species classification, large-scale natural image classification, attribute recognition and landmark localization.
Journal Title
IEEE Transactions on Artificial Intelligence
Conference Title
Book Title
Edition
Volume
Issue
Thesis Type
Degree Program
School
Publisher link
Patent number
Funder(s)
Grant identifier(s)
Rights Statement
Rights Statement
Item Access Status
Note
This publication has been entered in Griffith Research Online as an advanced online version.
Access the data
Related item(s)
Subject
Artificial intelligence
Deep learning
Computer vision and multimedia computation
Machine learning
Persistent link to this record
Citation
Yu, X; Zhao, Y; Gao, Y; Xiong, S, Learning Deep Asymmetric Tolerant Part Representation, IEEE Transactions on Artificial Intelligence, 2022