SlideShare a Scribd company logo
1
Click to edit Master title style
Zero-shot Image Recognition Using Relational
Matching, Adaptation and Calibration
Debasmit Das C.S. George Lee
Assistive Robotics Technology Laboratory
School of Electrical and Computer Engineering
Purdue University, West Lafayette, IN, USA
Funding Source : National Science Foundation (IIS-1813935), NVIDIA Hardware Grant
2
Outline
• INTRODUCTION
- Problem Description.
- Previous Work.
- Challenges.
• PROPOSED APPROACH
- Relational Matching.
- Domain Adaptation.
- Scaled Calibration.
• EXPERIMENTAL RESULTS
- Comparative studies.
- Parameter, Convergence results etc.
3
IntroductionZero Shot Learning (ZSL)
Feature Space
Semantic Space
• Base Categories (source domain) contain
abundant labeled data.
• Novel Categories (target domain) contain
unlabeled data.
• However, class level semantic information available
for all categories.
• Find relationship between feature and semantic
space.
Example
Target Domain
Source Domain
4
IntroductionRelated Work of ZSL
Zero-shot
Learning
Embedding
Methods
Transductive
approaches
Generative
approaches
Hybrid
approaches
• Linear embedding
[Bernardino et al. ICML’15]
• Deep Embedding
[Zhang et al. CVPR’17]
• Multiview
[Fu et al. TPAMI’15]
• Dictionary Learning
[Kodirov et al. ICCV’15]
• Constrained VAE
[Verma et al. CVPR’18]
• Feature GAN
[Xian et al. CVPR’18]
• Semantic Similarity
[Zhang et al. CVPR ’15]
• Convex Combo
[Norouzi et al. ICLR’13]
[Relate feature & semantics ]
[Use unlabeled test data] [Generate data]
[Novel class from old class]
5
Challenges of ZSL
Hubness Domain Shift Seen Class Biasedness
• In the GZSL Setting ,
test data can be from
both seen and
unseen categories.
• Most unseen test
data predicted as
seen categories.
• Initially studied by
Chao et al. ECCV’16.
• Domain shift between
unseen test data and
unseen semantic
embeddings.
• Since unseen test data
not used in training.
• Phenomenon where only
a few candidates become
nearest neighbor
predictions.
• Due to curse of
dimensionality.
• Initially studied by
Radovanovic et al.
JMLR’10.
Introduction
6
Proposed Solution
One-to-one and pairwise
regression
Domain Adaptation Calibration
• Need to adapt semantic
embeddings to unseen
test data.
• Use previous DA
approach [Das & Lee
EAAI’18].
• Find correspondences
between semantic
embedding and unseen
test samples.
• Scaled calibration to
reduce scores of seen
classes.
• Implicit reduction of
variance of seen
classes.
• Structural matching
between semantics
and feature.
• Implicit
reduction of
dimensionality.
Proposed Approach
ADDRESS HUBNESS ADDRESS DOMAIN
SHIFT
ADDRESS BIASEDNESS
7
Proposed Framework Proposed Approach
8
Relational Matching
• Firstly, match between a seen sample and the
corresponding semantic embedding.
• Secondly, try to match the structure (pair-wise
distance matrix) between the seen prototypes and
semantic embeddings.
One-to-one regression Pairwise regression
Minimize with gradient descent
Proposed Approach
9
Domain Adaptation
• Adapt the unseen semantic embeddings (A) close to
the unseen test data (U).
• Find correspondences (C) between each data point
and semantic embedding with class regularization.
Correspondence based loss Group Lasso based regularization
Conditional Gradient based optimization
Proposed Approach
10
Scaled Calibration
• Modify the nearest neighbor Euclidean distance scores.
• Euclidean distance scores for seen classes are scaled while
that of unseen classes are kept the same.
Seen Total
Unseen
Proposed Approach
11
Experimental Results
• Animals with Attributes (AwA2)
[Lampert et al. TPAMI’14]
(Att – 85, Ysrc - 40 , Ytar - 10 )
• Pascal & Yahoo (aPY)
[Farhadi et al. CVPR’09]
(Att – 64, Ysrc - 20 , Ytar - 12 )
• Caltech-UCSD Birds (CUB)
[Welinder et al. ‘10]
(Att – 312, Ysrc - 150 , Ytar - 50 )
• Scene Understanding (SUN)
[Patterson et al. CVPR’12]
(Att – 102, Ysrc - 645, Ytar - 72 )
DatasetsComparison with previous work on four datasets.
Comparative Study
tr – Unseen class accuracy in traditional setting
u – Unseen class accuracy in generalized setting
s – Seen class accuracy in generalized setting
H – Harmonic mean of u and s
R – Relational Matching
RA – Relational Matching + Domain Adaptation
RC – Relational Matching + Scaled Calibration
RAC – Relational Matching + Domain
Adaptation + Scaled Calibration
12
Experimental ResultsSensitivity Studies I
Effect of the calibration factor
Effect of the structural matching weight
13
Experimental ResultsSensitivity Studies II
Effect of changing the
proportion of seen classes
Effect of changing the
number of test samples
AwA2 SUN
14
Experimental ResultsConvergence Analysis
Convergence results on
AwA2 dataset
Convergence results on
SUN datasetEffect of no. of epochs
on test accuracy
15
Experimental ResultsVisualization & Hubness
Feature Visualization
Without Domain
Adaptation
With Domain
Adaptation
Hubness Measurement
Hubness measured using
skewness of NN
prediction disitribution.
Unseen Features Seen Features Unseen Semantic Embedding
Seen Semantic Embedding
16
• Three-step approach to ZSL with structural
matching, domain adaptation and calibration.
• Tested on four challenging ZSL datasets on which it
has substantial improvement in performance.
• Domain adaptation found to be most effective.
Hubness is also reduced.
Conclusion
Future Work
Distinguishing between novel and base categories and
investigate generative models.
17
THANK YOU
Any Questions ?

More Related Content

PDF
PhD Defense Slides
PDF
Preliminary Exam Slides
PPTX
Query dependent ranking using k nearest neighbor
 
PDF
Maximizing the Representation Gap between In-domain & OOD examples
PDF
Recommendation algorithm using reinforcement learning
PDF
Generative Models for General Audiences
PDF
Predicting best classifier using properties of data sets
PPTX
ICFHR 2014 Competition on Handwritten KeyWord Spotting (H-KWS 2014)
PhD Defense Slides
Preliminary Exam Slides
Query dependent ranking using k nearest neighbor
 
Maximizing the Representation Gap between In-domain & OOD examples
Recommendation algorithm using reinforcement learning
Generative Models for General Audiences
Predicting best classifier using properties of data sets
ICFHR 2014 Competition on Handwritten KeyWord Spotting (H-KWS 2014)

What's hot (20)

PDF
2019 dynamically composing_domain-data_selection_with_clean-data_selection_by...
PDF
Learning loss for active learning
PDF
Deep Learning for Natural Language Processing
PDF
Neural Semi-supervised Learning under Domain Shift
PDF
Chinese Named Entity Recognition with Graph-based Semi-supervised Learning Model
PDF
Research Inventy : International Journal of Engineering and Science
PPTX
A Collaborative Document Ranking Model for a Multi-Faceted Search
PPTX
Tanvi Motwani- A Few Examples Go A Long Way
PDF
JIST2015-Computing the Semantic Similarity of Resources in DBpedia for Recomm...
PPTX
JIST2015-data challenge
PDF
PR-258: From ImageNet to Image Classification: Contextualizing Progress on Be...
PDF
Transfer Learning -- The Next Frontier for Machine Learning
PPTX
A Study on the Importance of Adaptive Seed Value Exploration
PDF
Improving neural question generation using answer separation
PDF
Artificial Intelligence Course: Linear models
PDF
BOIL: Towards Representation Change for Few-shot Learning
PDF
Large Scale GAN Training for High Fidelity Natural Image Synthesis
PPTX
One shot learning
PPTX
Talk@rmit 09112017
PDF
Machine Learning Lecture 3 Decision Trees
2019 dynamically composing_domain-data_selection_with_clean-data_selection_by...
Learning loss for active learning
Deep Learning for Natural Language Processing
Neural Semi-supervised Learning under Domain Shift
Chinese Named Entity Recognition with Graph-based Semi-supervised Learning Model
Research Inventy : International Journal of Engineering and Science
A Collaborative Document Ranking Model for a Multi-Faceted Search
Tanvi Motwani- A Few Examples Go A Long Way
JIST2015-Computing the Semantic Similarity of Resources in DBpedia for Recomm...
JIST2015-data challenge
PR-258: From ImageNet to Image Classification: Contextualizing Progress on Be...
Transfer Learning -- The Next Frontier for Machine Learning
A Study on the Importance of Adaptive Seed Value Exploration
Improving neural question generation using answer separation
Artificial Intelligence Course: Linear models
BOIL: Towards Representation Change for Few-shot Learning
Large Scale GAN Training for High Fidelity Natural Image Synthesis
One shot learning
Talk@rmit 09112017
Machine Learning Lecture 3 Decision Trees
Ad

Similar to Zero-shot Image Recognition Using Relational Matching, Adaptation and Calibration (20)

PPTX
STRICT: Information Retrieval Based Search Term Identification for Concept Lo...
PPTX
Surprise Adequacy for Deep Learning Systems (SADL)
PPTX
Graph Matching Unsupervised Domain Adaptation
PPTX
Ensemble_instance_unsupersied_learning 01_02_2024.pptx
PPT
probabilistic ranking
PDF
Icann2018ppt final
PPTX
Smoothed manifold
PDF
Towards Set Learning and Prediction - Laura Leal-Taixe - UPC Barcelona 2018
PDF
Seminar - Similarity Joins in SQL (performance and semantic joins)
PDF
Lecture-2-Relational-Algebra-and-SQL-Advanced-DataBase-Theory-MS.pdf
PPTX
240408_JW_labseminar[Asymmetric Transitivity Preserving Graph Embedding].pptx
PPTX
NS-CUK Seminar: J.H.Lee, Review on "Relational Self-Supervised Learning on Gr...
PPT
Part 1
PDF
Natural language processing for extracting synthesis recipes and applications...
PPTX
0 introduction
PPT
about data mining and Exp about data mining and Exp.
PPTX
Compressing Graphs and Indexes with Recursive Graph Bisection
PDF
Presentation File of paper "Leveraging Normalization Layer in Adapters With P...
PDF
Data-analytic sins in property-based molecular design
PDF
How much position information do convolutional neural networks encode? review...
STRICT: Information Retrieval Based Search Term Identification for Concept Lo...
Surprise Adequacy for Deep Learning Systems (SADL)
Graph Matching Unsupervised Domain Adaptation
Ensemble_instance_unsupersied_learning 01_02_2024.pptx
probabilistic ranking
Icann2018ppt final
Smoothed manifold
Towards Set Learning and Prediction - Laura Leal-Taixe - UPC Barcelona 2018
Seminar - Similarity Joins in SQL (performance and semantic joins)
Lecture-2-Relational-Algebra-and-SQL-Advanced-DataBase-Theory-MS.pdf
240408_JW_labseminar[Asymmetric Transitivity Preserving Graph Embedding].pptx
NS-CUK Seminar: J.H.Lee, Review on "Relational Self-Supervised Learning on Gr...
Part 1
Natural language processing for extracting synthesis recipes and applications...
0 introduction
about data mining and Exp about data mining and Exp.
Compressing Graphs and Indexes with Recursive Graph Bisection
Presentation File of paper "Leveraging Normalization Layer in Adapters With P...
Data-analytic sins in property-based molecular design
How much position information do convolutional neural networks encode? review...
Ad

More from Debasmit Das (7)

PDF
Icip2018posterv3
PPTX
Recursive Oscillators
PPTX
Surface Plasmon Resonance
PPTX
IIT Roorkee Motorsports (IITRMS)
PPT
Pseudo Spectral Optimal Control for Coverage Path Planning
PPTX
Intention Inference
PPT
Resistive Sensors
Icip2018posterv3
Recursive Oscillators
Surface Plasmon Resonance
IIT Roorkee Motorsports (IITRMS)
Pseudo Spectral Optimal Control for Coverage Path Planning
Intention Inference
Resistive Sensors

Recently uploaded (20)

PPT
Miokarditis (Inflamasi pada Otot Jantung)
PDF
annual-report-2024-2025 original latest.
PPTX
Introduction to Basics of Ethical Hacking and Penetration Testing -Unit No. 1...
PPTX
iec ppt-1 pptx icmr ppt on rehabilitation.pptx
PPTX
Microsoft-Fabric-Unifying-Analytics-for-the-Modern-Enterprise Solution.pptx
PDF
Foundation of Data Science unit number two notes
PPT
Reliability_Chapter_ presentation 1221.5784
PPTX
MODULE 8 - DISASTER risk PREPAREDNESS.pptx
PPTX
Introduction-to-Cloud-ComputingFinal.pptx
PPT
Quality review (1)_presentation of this 21
PDF
168300704-gasification-ppt.pdfhghhhsjsjhsuxush
PPTX
Computer network topology notes for revision
PDF
Business Analytics and business intelligence.pdf
PPTX
AI Strategy room jwfjksfksfjsjsjsjsjfsjfsj
PDF
BF and FI - Blockchain, fintech and Financial Innovation Lesson 2.pdf
PPTX
Business Ppt On Nestle.pptx huunnnhhgfvu
PPTX
oil_refinery_comprehensive_20250804084928 (1).pptx
PPTX
Introduction to Knowledge Engineering Part 1
PDF
“Getting Started with Data Analytics Using R – Concepts, Tools & Case Studies”
PDF
Lecture1 pattern recognition............
Miokarditis (Inflamasi pada Otot Jantung)
annual-report-2024-2025 original latest.
Introduction to Basics of Ethical Hacking and Penetration Testing -Unit No. 1...
iec ppt-1 pptx icmr ppt on rehabilitation.pptx
Microsoft-Fabric-Unifying-Analytics-for-the-Modern-Enterprise Solution.pptx
Foundation of Data Science unit number two notes
Reliability_Chapter_ presentation 1221.5784
MODULE 8 - DISASTER risk PREPAREDNESS.pptx
Introduction-to-Cloud-ComputingFinal.pptx
Quality review (1)_presentation of this 21
168300704-gasification-ppt.pdfhghhhsjsjhsuxush
Computer network topology notes for revision
Business Analytics and business intelligence.pdf
AI Strategy room jwfjksfksfjsjsjsjsjfsjfsj
BF and FI - Blockchain, fintech and Financial Innovation Lesson 2.pdf
Business Ppt On Nestle.pptx huunnnhhgfvu
oil_refinery_comprehensive_20250804084928 (1).pptx
Introduction to Knowledge Engineering Part 1
“Getting Started with Data Analytics Using R – Concepts, Tools & Case Studies”
Lecture1 pattern recognition............

Zero-shot Image Recognition Using Relational Matching, Adaptation and Calibration

  • 1. 1 Click to edit Master title style Zero-shot Image Recognition Using Relational Matching, Adaptation and Calibration Debasmit Das C.S. George Lee Assistive Robotics Technology Laboratory School of Electrical and Computer Engineering Purdue University, West Lafayette, IN, USA Funding Source : National Science Foundation (IIS-1813935), NVIDIA Hardware Grant
  • 2. 2 Outline • INTRODUCTION - Problem Description. - Previous Work. - Challenges. • PROPOSED APPROACH - Relational Matching. - Domain Adaptation. - Scaled Calibration. • EXPERIMENTAL RESULTS - Comparative studies. - Parameter, Convergence results etc.
  • 3. 3 IntroductionZero Shot Learning (ZSL) Feature Space Semantic Space • Base Categories (source domain) contain abundant labeled data. • Novel Categories (target domain) contain unlabeled data. • However, class level semantic information available for all categories. • Find relationship between feature and semantic space. Example Target Domain Source Domain
  • 4. 4 IntroductionRelated Work of ZSL Zero-shot Learning Embedding Methods Transductive approaches Generative approaches Hybrid approaches • Linear embedding [Bernardino et al. ICML’15] • Deep Embedding [Zhang et al. CVPR’17] • Multiview [Fu et al. TPAMI’15] • Dictionary Learning [Kodirov et al. ICCV’15] • Constrained VAE [Verma et al. CVPR’18] • Feature GAN [Xian et al. CVPR’18] • Semantic Similarity [Zhang et al. CVPR ’15] • Convex Combo [Norouzi et al. ICLR’13] [Relate feature & semantics ] [Use unlabeled test data] [Generate data] [Novel class from old class]
  • 5. 5 Challenges of ZSL Hubness Domain Shift Seen Class Biasedness • In the GZSL Setting , test data can be from both seen and unseen categories. • Most unseen test data predicted as seen categories. • Initially studied by Chao et al. ECCV’16. • Domain shift between unseen test data and unseen semantic embeddings. • Since unseen test data not used in training. • Phenomenon where only a few candidates become nearest neighbor predictions. • Due to curse of dimensionality. • Initially studied by Radovanovic et al. JMLR’10. Introduction
  • 6. 6 Proposed Solution One-to-one and pairwise regression Domain Adaptation Calibration • Need to adapt semantic embeddings to unseen test data. • Use previous DA approach [Das & Lee EAAI’18]. • Find correspondences between semantic embedding and unseen test samples. • Scaled calibration to reduce scores of seen classes. • Implicit reduction of variance of seen classes. • Structural matching between semantics and feature. • Implicit reduction of dimensionality. Proposed Approach ADDRESS HUBNESS ADDRESS DOMAIN SHIFT ADDRESS BIASEDNESS
  • 8. 8 Relational Matching • Firstly, match between a seen sample and the corresponding semantic embedding. • Secondly, try to match the structure (pair-wise distance matrix) between the seen prototypes and semantic embeddings. One-to-one regression Pairwise regression Minimize with gradient descent Proposed Approach
  • 9. 9 Domain Adaptation • Adapt the unseen semantic embeddings (A) close to the unseen test data (U). • Find correspondences (C) between each data point and semantic embedding with class regularization. Correspondence based loss Group Lasso based regularization Conditional Gradient based optimization Proposed Approach
  • 10. 10 Scaled Calibration • Modify the nearest neighbor Euclidean distance scores. • Euclidean distance scores for seen classes are scaled while that of unseen classes are kept the same. Seen Total Unseen Proposed Approach
  • 11. 11 Experimental Results • Animals with Attributes (AwA2) [Lampert et al. TPAMI’14] (Att – 85, Ysrc - 40 , Ytar - 10 ) • Pascal & Yahoo (aPY) [Farhadi et al. CVPR’09] (Att – 64, Ysrc - 20 , Ytar - 12 ) • Caltech-UCSD Birds (CUB) [Welinder et al. ‘10] (Att – 312, Ysrc - 150 , Ytar - 50 ) • Scene Understanding (SUN) [Patterson et al. CVPR’12] (Att – 102, Ysrc - 645, Ytar - 72 ) DatasetsComparison with previous work on four datasets. Comparative Study tr – Unseen class accuracy in traditional setting u – Unseen class accuracy in generalized setting s – Seen class accuracy in generalized setting H – Harmonic mean of u and s R – Relational Matching RA – Relational Matching + Domain Adaptation RC – Relational Matching + Scaled Calibration RAC – Relational Matching + Domain Adaptation + Scaled Calibration
  • 12. 12 Experimental ResultsSensitivity Studies I Effect of the calibration factor Effect of the structural matching weight
  • 13. 13 Experimental ResultsSensitivity Studies II Effect of changing the proportion of seen classes Effect of changing the number of test samples AwA2 SUN
  • 14. 14 Experimental ResultsConvergence Analysis Convergence results on AwA2 dataset Convergence results on SUN datasetEffect of no. of epochs on test accuracy
  • 15. 15 Experimental ResultsVisualization & Hubness Feature Visualization Without Domain Adaptation With Domain Adaptation Hubness Measurement Hubness measured using skewness of NN prediction disitribution. Unseen Features Seen Features Unseen Semantic Embedding Seen Semantic Embedding
  • 16. 16 • Three-step approach to ZSL with structural matching, domain adaptation and calibration. • Tested on four challenging ZSL datasets on which it has substantial improvement in performance. • Domain adaptation found to be most effective. Hubness is also reduced. Conclusion Future Work Distinguishing between novel and base categories and investigate generative models.

Editor's Notes

  • #3: Introduction – introduction to the problem and previous work Our method – problem formulation, optimization and proposed solution Experimental results – experimental results and discussions
  • #4: Give an example of a classification setting
  • #5: Just talk about the limitations of these methods and no more details
  • #6: Talk in details and study about hubness.
  • #7: COMBINATION OF EMBEDDING AND TRANSDUCTIVE EXCEPT THAT POST-PROCESSING INSTEAD OF DIRECT TRANSDUCTIVE IS USED
  • #8: 3 step procedure
  • #9: Talk that local method may have better accuracy but it might be slower
  • #10: Talk that local method may have better accuracy but it might be slower
  • #11: Talk that local method may have better accuracy but it might be slower
  • #12: Explain
  • #13: Explain
  • #14: Explain
  • #15: Explain
  • #16: Explain