Indirect visual–semantic alignment for generalized zero-shot recognition

Yan He Chen, Mei Chen Yeh*

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

Our paper addresses the challenge of generalized zero-shot learning, where the label of a target image may belong to either a seen or an unseen category. Previous methods for this task typically learn a joint embedding space where image features and their corresponding class prototypes are directly aligned. However, this can be difficult due to the inherent gap between the visual and semantic space. To overcome this challenge, we propose a novel learning framework that relaxes the alignment requirement. Our approach employs a metric learning-based loss function to optimize the visual embedding model, allowing for different penalty strengths on within-class and between-class similarities. By avoiding pair-wise comparisons between image and class embeddings, our approach achieves more flexibility in learning discriminative and generalized visual features. Our extensive experiments demonstrate the superiority of our method with performance on par with the state-of-the-art on five benchmarks.

Original languageEnglish
Article number111
JournalMultimedia Systems
Volume30
Issue number2
DOIs
Publication statusPublished - 2024 Apr

Keywords

  • Deep metric learning
  • Fine-grained visual recognition
  • Generalized zero-shot learning
  • Visual–semantic embedding

ASJC Scopus subject areas

  • Software
  • Information Systems
  • Media Technology
  • Hardware and Architecture
  • Computer Networks and Communications

Fingerprint

Dive into the research topics of 'Indirect visual–semantic alignment for generalized zero-shot recognition'. Together they form a unique fingerprint.

Cite this