SlideShare a Scribd company logo
Hierarchical Novelty Detection
for Visual Object Recognition
Kibok Lee*, Kimin Lee†, Kyle Min*,
Yuting Zhang*, Jinwoo Shin†, Honglak Lee*‡
University of Michigan*, KAIST†, Google Brain‡
Why Hierarchical Novelty Detection?
• Conventional novelty detection framework does not provide more
information than “novelty” of an object.
• Suppose we have training data like …
Pomeranian Welsh corgiPersian cat Siamese cat
Why Hierarchical Novelty Detection?
• Then, suppose we have test images like …
Test image:
True label: Siamese cat Angora cat Dachshund Pika
Prior works: Siamese cat novel novel novel
Pomeranian Welsh corgiPersian cat Siamese cat
Why Hierarchical Novelty Detection?
Ours: Siamese cat novel cat novel dog novel animal
animal
dogcat
Pomeranian Welsh corgiPersian cat Siamese cat
True label: Siamese cat Angora cat Dachshund Pika
Test image:
Prior works: Siamese cat novel novel novel
Why Hierarchical Novelty Detection?
• Our hierarchical novelty detection framework aims to find the most
specific class label of any data on the hierarchical taxonomy built
with known labels.
Ours: Siamese cat novel cat novel dog novel animal
True label: Siamese cat Angora cat Dachshund Pika
Test image:
Prior works: Siamese cat novel novel novel
Why Hierarchical Novelty Detection?
• This framework can be potentially useful for automatically or
interactively organizing a customized taxonomy
• Company’s product catalog
• Wildlife monitoring
• Personal photo library
• by suggesting closest categories for an image from novel
categories.
• new consumer products
• unregistered animal species
• untagged scenes or places
Hierarchical Taxonomy
• Our taxonomy has three types of classes.
• Known leaf classes are seen during training
• Super classes are ancestors of leaf classes, also known
• Novel classes are unseen during training
• Their expected prediction is the closest super class in the taxonomy in our task
Hierarchical Taxonomy
Taxonomy of known classes Set of all known leaf classes
Set of parents Set of novel classes whose closest known class is
Set of children Set of known classes except and its descendants
Set of ancestors including itself:
Hierarchical Taxonomy
Taxonomy of known classes Set of all known leaf classes
Set of parents Set of novel classes whose closest known class is
Set of children Set of known classes except and its descendants
Set of ancestors including itself:
Hierarchical Taxonomy
Taxonomy of known classes Set of all known leaf classes
Set of parents Set of novel classes whose closest known class is
Set of children Set of known classes except and its descendants
Set of ancestors including itself:
Hierarchical Taxonomy
Taxonomy of known classes Set of all known leaf classes
Set of parents Set of novel classes whose closest known class is
Set of children Set of known classes except and its descendants
Set of ancestors including itself:
Hierarchical Taxonomy
Taxonomy of known classes Set of all known leaf classes
Set of parents Set of novel classes whose closest known class is
Set of children Set of known classes except and its descendants
Set of ancestors including itself:
Hierarchical Taxonomy
Taxonomy of known classes Set of all known leaf classes
Set of parents Set of novel classes whose closest known class is
Set of children Set of known classes except and its descendants
Set of ancestors including itself:
Approach - Top-down (TD) Method
• Multi-stage classification
• Until arriving at a known leaf class
.2 .8
.7 .1 .2
Approach - Top-down (TD) Method
• Multi-stage classification
• Until arriving at a known leaf class
• Or classification is unconfident (novel)
.5 .5
Approach - Top-down (TD) Method
• Classification rule:
• Classification is confident if
Approach - Top-down (TD) Method
• Training objective:
Approach - Flatten Method
• Represent all probabilities of known leaf and novel classes in a
single vector
• Add virtual novel classes
Approach - Flatten Method
• Represent all probabilities of known leaf and novel classes in a
single vector
• Add virtual novel classes
• And then flatten the structure
Approach - Flatten Method
• Represent all probabilities of known leaf and novel classes in a
single vector
• Add virtual novel classes
• And then flatten the structure
Approach - Flatten Method
• Classification rule:
• We propose two strategies to train this model.
Approach - Flatten Method
• Data relabeling
• Fill novel classes by hierarchical relabeling
animal
cat dog
Persian cat Siamese cat Pomeranian Welsh corgi
Approach - Flatten Method
• Data relabeling
• Fill novel classes by hierarchical relabeling
animal
cat dog
Persian cat Siamese cat Pomeranian Welsh corgi
Approach - Flatten Method
• Data relabeling
• Fill novel classes by hierarchical relabeling
animal
cat dog
Persian cat Siamese cat Pomeranian Welsh corgi
Approach - Flatten Method
• Data relabeling
• Fill novel classes by hierarchical relabeling
• Relabeling rate can be chosen by validation
animal
cat dog
Persian cat Siamese cat Pomeranian Welsh corgi
Approach - Flatten Method
• Data relabeling
• Training objective:
animal
cat dog
Persian cat Siamese cat Pomeranian Welsh corgi
Approach - Flatten Method
• Leave-one-out (LOO) strategy
• Generate deficient taxonomies
• And then train the model with them
animal
cat dog
Persian cat Siamese cat Pomeranian Welsh corgi
Approach - Flatten Method
• Leave-one-out (LOO) strategy
• e.g., when
animal
cat dog
Persian cat Siamese cat Pomeranian Welsh corgi
Approach - Flatten Method
• Leave-one-out (LOO) strategy
• e.g., when
animal
novel animal dog
Pomeranian Welsh corgi
Approach - Flatten Method
• Leave-one-out (LOO) strategy
• Training objective:
animal
novel animal dog
Pomeranian Welsh corgi
Approach - Combined Method (TD+LOO)
• Compute and enumerate the output of TD
• And then feed it to LOO
• Combined method utilizes their complementary benefits.
• Top-down method leverages hierarchical structure information.
• But it suffers from error aggregation over hierarchy.
• Flatten method avoids error aggregation over hierarchy.
• But it does not leverage hierarchical structure information.
Experiments - Hierarchical Novelty Detection
• Quantitative results
Method
ImageNet AwA2 CUB
Novel AUC Novel AUC Novel AUC
DARTS 10.89 8.83 36.75 35.14 40.42 30.07
Relabel 15.29 11.51 45.71 40.28 38.23 28.75
LOO 15.72 12.00 50.00 43.63 40.78 31.92
TD+LOO 18.78 13.98 53.57 46.77 43.29 33.16
Experiments - Hierarchical Novelty Detection
• Compared algorithms
• Baseline: DARTS (Deng et al., 2012)
• Ours: Relabel, LOO, TD+LOO
Method
ImageNet AwA2 CUB
Novel AUC Novel AUC Novel AUC
DARTS 10.89 8.83 36.75 35.14 40.42 30.07
Relabel 15.29 11.51 45.71 40.28 38.23 28.75
LOO 15.72 12.00 50.00 43.63 40.78 31.92
TD+LOO 18.78 13.98 53.57 46.77 43.29 33.16
J. Deng, J. Krause, A. C. Berg, and L. Fei-Fei. “Hedging your bets: Optimizing accuracy-specificity trade-offs in large scale visual recognition.” In CVPR, 2012
Experiments - Hierarchical Novelty Detection
• Datasets
• ImageNet: 1k known, 16k novel classes
• AwA2: 40 known, 10 novel classes
• CUB: 150 known, 50 novel classes
Method
ImageNet AwA2 CUB
Novel AUC Novel AUC Novel AUC
DARTS 10.89 8.83 36.75 35.14 40.42 30.07
Relabel 15.29 11.51 45.71 40.28 38.23 28.75
LOO 15.72 12.00 50.00 43.63 40.78 31.92
TD+LOO 18.78 13.98 53.57 46.77 43.29 33.16
Experiments - Hierarchical Novelty Detection
• Metrics
• Novel class accuracy @ known class accuracy = 50%
• By adding an appropriate score bias to all novel classes
• Area under known-novel class accuracy curve
• By varying the novel class score bias
Method
ImageNet AwA2 CUB
Novel AUC Novel AUC Novel AUC
DARTS 10.89 8.83 36.75 35.14 40.42 30.07
Relabel 15.29 11.51 45.71 40.28 38.23 28.75
LOO 15.72 12.00 50.00 43.63 40.78 31.92
TD+LOO 18.78 13.98 53.57 46.77 43.29 33.16
Experiments - Hierarchical Novelty Detection
• Metrics
• Novel class accuracy @ known class accuracy = 50%
• By adding an appropriate score bias to all novel classes
• Area under known-novel class accuracy curve
• By varying the novel class score bias
(a) ImageNet (b) AwA2 (c) CUB
Experiments - Hierarchical Novelty Detection
• Metrics
• Novel class accuracy @ known class accuracy = 50%
• By adding an appropriate score bias to all novel classes
• Area under known-novel class accuracy curve
• By varying the novel class score bias
(a) ImageNet (b) AwA2 (c) CUB
Experiments - Hierarchical Novelty Detection
• Metrics
• Novel class accuracy @ known class accuracy = 50%
• By adding an appropriate score bias to all novel classes
• Area under known-novel class accuracy curve
• By varying the novel class score bias
(a) ImageNet (b) AwA2 (c) CUB
Experiments - Hierarchical Novelty Detection
• Qualitative results
Novel class: American foxhound
GT foxhound
Experiments - Hierarchical Novelty Detection
• Qualitative results
Novel class: American foxhound
Method ε A Word
GT foxhound
DARTS 2 N beagle
Relabel 1 Y hound dog
LOO 0 Y foxhound
TD+LOO 0 Y foxhound
Experiments - Hierarchical Novelty Detection
• Qualitative results
Novel class: serval
Method ε A Word
GT wildcat
DARTS 3 N Egyptian cat
Relabel 2 N domestic cat
LOO 2 Y feline
TD+LOO 1 Y cat
Experiments - Hierarchical Novelty Detection
• Qualitative results
Novel class: song thrush
Method ε A Word
GT thrush
DARTS 3 N hummingbird
Relabel 2 Y bird
LOO 1 Y oscine bird
TD+LOO 0 Y thrush
Experiments - Hierarchical Novelty Detection
• Qualitative results
Novel class: ice-cream sundae
Method ε A Word
GT frozen dessert
DARTS 4 Y food, nutrient
Relabel 1 N ice cream
LOO 1 Y dessert
TD+LOO 0 Y frozen dessert
Experiments - Generalized Zero-Shot Learning
• Quantitative results
Embedding AwA1 AwA2 CUB
Att Word Hier Unseen AUC Unseen AUC Unseen AUC
√ 65.29 50.02 63.87 51.27 50.05 23.60
√ 51.87 39.67 54.77 42.21 27.28 11.47
√ √ 67.80 52.84 65.76 53.18 49.83 24.13
Path 42.57 30.58 44.34 33.44 24.22 8.38
√ Path 67.09 51.45 66.58 53.50 50.25 23.70
√ Path 52.89 40.66 55.28 42.86 27.72 11.65
√ √ Path 68.04 53.21 67.28 54.31 50.87 24.20
TD 33.86 25.56 31.84 24.97 13.09 7.20
√ TD 66.13 54.66 66.86 57.49 50.17 30.31
√ TD 56.14 46.28 59.67 49.39 29.05 16.73
√ √ TD 69.23 57.67 68.80 59.24 50.17 30.31
Experiments - Generalized Zero-Shot Learning
• Semantic embeddings
• Attributes (numeric attribute values)
• Word vector (similarity among words in real coordinate space)
• Hierarchical embedding
Embedding AwA1 AwA2 CUB
Att Word Hier Unseen AUC Unseen AUC Unseen AUC
√ 65.29 50.02 63.87 51.27 50.05 23.60
√ 51.87 39.67 54.77 42.21 27.28 11.47
√ √ 67.80 52.84 65.76 53.18 49.83 24.13
Path 42.57 30.58 44.34 33.44 24.22 8.38
√ Path 67.09 51.45 66.58 53.50 50.25 23.70
√ Path 52.89 40.66 55.28 42.86 27.72 11.65
√ √ Path 68.04 53.21 67.28 54.31 50.87 24.20
TD 33.86 25.56 31.84 24.97 13.09 7.20
√ TD 66.13 54.66 66.86 57.49 50.17 30.31
√ TD 56.14 46.28 59.67 49.39 29.05 16.73
√ √ TD 69.23 57.67 68.80 59.24 50.17 30.31
Experiments - Generalized Zero-Shot Learning
• Compared hierarchical embeddings
• Baseline: Path (Akata et al., 2015)
• Distance between classes on hierarchy
• Ours: Top-down (TD)
• Expected output of our top-down model
Embedding AwA1 AwA2 CUB
Att Word Hier Unseen AUC Unseen AUC Unseen AUC
√ 65.29 50.02 63.87 51.27 50.05 23.60
√ 51.87 39.67 54.77 42.21 27.28 11.47
√ √ 67.80 52.84 65.76 53.18 49.83 24.13
Path 42.57 30.58 44.34 33.44 24.22 8.38
√ Path 67.09 51.45 66.58 53.50 50.25 23.70
√ Path 52.89 40.66 55.28 42.86 27.72 11.65
√ √ Path 68.04 53.21 67.28 54.31 50.87 24.20
TD 33.86 25.56 31.84 24.97 13.09 7.20
√ TD 66.13 54.66 66.86 57.49 50.17 30.31
√ TD 56.14 46.28 59.67 49.39 29.05 16.73
√ √ TD 69.23 57.67 68.80 59.24 50.17 30.31
Z. Akata, S. Reed, D. Walter, H. Lee, and B. Schiele. “Evaluation of output embeddings for fine-grained image classification.” In CVPR, 2015.
Experiments - Generalized Zero-Shot Learning
• Datasets
• AwA1,2: 40 known, 10 novel classes
• CUB: 150 known, 50 novel classes
Embedding AwA1 AwA2 CUB
Att Word Hier Unseen AUC Unseen AUC Unseen AUC
√ 65.29 50.02 63.87 51.27 50.05 23.60
√ 51.87 39.67 54.77 42.21 27.28 11.47
√ √ 67.80 52.84 65.76 53.18 49.83 24.13
Path 42.57 30.58 44.34 33.44 24.22 8.38
√ Path 67.09 51.45 66.58 53.50 50.25 23.70
√ Path 52.89 40.66 55.28 42.86 27.72 11.65
√ √ Path 68.04 53.21 67.28 54.31 50.87 24.20
TD 33.86 25.56 31.84 24.97 13.09 7.20
√ TD 66.13 54.66 66.86 57.49 50.17 30.31
√ TD 56.14 46.28 59.67 49.39 29.05 16.73
√ √ TD 69.23 57.67 68.80 59.24 50.17 30.31
Experiments - Generalized Zero-Shot Learning
• Metrics
• Unseen class accuracy (ZSL)
• Area under seen-unseen curve (GZSL)
• By varying the unseen class score bias
Embedding AwA1 AwA2 CUB
Att Word Hier Unseen AUC Unseen AUC Unseen AUC
√ 65.29 50.02 63.87 51.27 50.05 23.60
√ 51.87 39.67 54.77 42.21 27.28 11.47
√ √ 67.80 52.84 65.76 53.18 49.83 24.13
Path 42.57 30.58 44.34 33.44 24.22 8.38
√ Path 67.09 51.45 66.58 53.50 50.25 23.70
√ Path 52.89 40.66 55.28 42.86 27.72 11.65
√ √ Path 68.04 53.21 67.28 54.31 50.87 24.20
TD 33.86 25.56 31.84 24.97 13.09 7.20
√ TD 66.13 54.66 66.86 57.49 50.17 30.31
√ TD 56.14 46.28 59.67 49.39 29.05 16.73
√ √ TD 69.23 57.67 68.80 59.24 50.17 30.31
Experiments - Generalized Zero-Shot Learning
• Metrics
• Unseen class accuracy (ZSL)
• Area under seen-unseen curve (GZSL)
• By varying the unseen class score bias
(a) AwA1 (b) AwA2 (c) CUB
Experiments - Generalized Zero-Shot Learning
• Metrics
• Unseen class accuracy (ZSL)
• Area under seen-unseen curve (GZSL)
• By varying the unseen class score bias
(a) AwA1 (b) AwA2 (c) CUB
Experiments - Generalized Zero-Shot Learning
• Metrics
• Unseen class accuracy (ZSL)
• Area under seen-unseen curve (GZSL)
• By varying the unseen class score bias
(a) AwA1 (b) AwA2 (c) CUB
Conclusion
• We propose a hierarchical novelty detection framework, which
aims to find the most specific class label of any data on the
hierarchical taxonomy built with known labels.
• We propose two approaches.
• Top-down method performs classification & novelty detection
hierarchically.
• Flatten method computes single probability vector of all candidate classes.
• Their combination takes advantage of their complementary benefits.
• Our model can be combined with other semantic embeddings to
improve generalized zero-shot learning performance.

More Related Content

PPTX
Knowing what we’re talking about
PDF
Machine Learning - Classification (ctd.)
PDF
Link Discovery Tutorial Part II: Accuracy
PDF
IRJET - Automating the Identification of Forest Animals and Alerting in Case ...
PDF
Similarity Features, and their Role in Concept Alignment Learning
PDF
ANIMAL SPECIES RECOGNITION SYSTEM USING DEEP LEARNING
PDF
Avoid Overfitting with Regularization
PDF
群衆の知を引き出すための機械学習(第4回ステアラボ人工知能セミナー)
Knowing what we’re talking about
Machine Learning - Classification (ctd.)
Link Discovery Tutorial Part II: Accuracy
IRJET - Automating the Identification of Forest Animals and Alerting in Case ...
Similarity Features, and their Role in Concept Alignment Learning
ANIMAL SPECIES RECOGNITION SYSTEM USING DEEP LEARNING
Avoid Overfitting with Regularization
群衆の知を引き出すための機械学習(第4回ステアラボ人工知能セミナー)

Similar to Hierarchical Novelty Detection for Visual Object Recognition (20)

PDF
Intelligent Multimedia Recommendation
PPT
lecture_mooney.ppt
PPTX
Self-ensembling for visual domain adpation
PPT
l15.ppt
PDF
Automated building of taxonomies for search engines
PDF
07 dimensionality reduction
PPTX
in5490-classification (1).pptx
PDF
Addressing overfitting in comparative study for deep learningbased classifica...
PDF
Artificial Intelligence: an introduction.pdf
PDF
02 - Data validation and validity deze keer
PDF
Image similarity with deep learning
PDF
Noorbehbahani a survey on instance selection for active learning
PDF
Novelty generation with deep learning
PDF
Grammarly AI-NLP Club #10 - Information-Theoretic Probing with Minimum Descri...
PPT
S10
PPT
S10
PDF
Icann2018ppt final
PDF
Modeling and Aggregation of Complex Annotations
PDF
Week 4 advanced labeling, augmentation and data preprocessing
PPT
Supervised_Learning.ppt
Intelligent Multimedia Recommendation
lecture_mooney.ppt
Self-ensembling for visual domain adpation
l15.ppt
Automated building of taxonomies for search engines
07 dimensionality reduction
in5490-classification (1).pptx
Addressing overfitting in comparative study for deep learningbased classifica...
Artificial Intelligence: an introduction.pdf
02 - Data validation and validity deze keer
Image similarity with deep learning
Noorbehbahani a survey on instance selection for active learning
Novelty generation with deep learning
Grammarly AI-NLP Club #10 - Information-Theoretic Probing with Minimum Descri...
S10
S10
Icann2018ppt final
Modeling and Aggregation of Complex Annotations
Week 4 advanced labeling, augmentation and data preprocessing
Supervised_Learning.ppt
Ad

More from NAVER Engineering (20)

PDF
React vac pattern
PDF
디자인 시스템에 직방 ZUIX
PDF
진화하는 디자인 시스템(걸음마 편)
PDF
서비스 운영을 위한 디자인시스템 프로젝트
PDF
BPL(Banksalad Product Language) 무야호
PDF
이번 생에 디자인 시스템은 처음이라
PDF
날고 있는 여러 비행기 넘나 들며 정비하기
PDF
쏘카프레임 구축 배경과 과정
PDF
플랫폼 디자이너 없이 디자인 시스템을 구축하는 프로덕트 디자이너의 우당탕탕 고통 연대기
PDF
200820 NAVER TECH CONCERT 15_Code Review is Horse(코드리뷰는 말이야)(feat.Latte)
PDF
200819 NAVER TECH CONCERT 03_화려한 코루틴이 내 앱을 감싸네! 코루틴으로 작성해보는 깔끔한 비동기 코드
PDF
200819 NAVER TECH CONCERT 10_맥북에서도 아이맥프로에서 빌드하는 것처럼 빌드 속도 빠르게 하기
PDF
200819 NAVER TECH CONCERT 08_성능을 고민하는 슬기로운 개발자 생활
PDF
200819 NAVER TECH CONCERT 05_모르면 손해보는 Android 디버깅/분석 꿀팁 대방출
PDF
200819 NAVER TECH CONCERT 09_Case.xcodeproj - 좋은 동료로 거듭나기 위한 노하우
PDF
200820 NAVER TECH CONCERT 14_야 너두 할 수 있어. 비전공자, COBOL 개발자를 거쳐 네이버에서 FE 개발하게 된...
PDF
200820 NAVER TECH CONCERT 13_네이버에서 오픈 소스 개발을 통해 성장하는 방법
PDF
200820 NAVER TECH CONCERT 12_상반기 네이버 인턴을 돌아보며
PDF
200820 NAVER TECH CONCERT 11_빠르게 성장하는 슈퍼루키로 거듭나기
PDF
200819 NAVER TECH CONCERT 07_신입 iOS 개발자 개발업무 적응기
React vac pattern
디자인 시스템에 직방 ZUIX
진화하는 디자인 시스템(걸음마 편)
서비스 운영을 위한 디자인시스템 프로젝트
BPL(Banksalad Product Language) 무야호
이번 생에 디자인 시스템은 처음이라
날고 있는 여러 비행기 넘나 들며 정비하기
쏘카프레임 구축 배경과 과정
플랫폼 디자이너 없이 디자인 시스템을 구축하는 프로덕트 디자이너의 우당탕탕 고통 연대기
200820 NAVER TECH CONCERT 15_Code Review is Horse(코드리뷰는 말이야)(feat.Latte)
200819 NAVER TECH CONCERT 03_화려한 코루틴이 내 앱을 감싸네! 코루틴으로 작성해보는 깔끔한 비동기 코드
200819 NAVER TECH CONCERT 10_맥북에서도 아이맥프로에서 빌드하는 것처럼 빌드 속도 빠르게 하기
200819 NAVER TECH CONCERT 08_성능을 고민하는 슬기로운 개발자 생활
200819 NAVER TECH CONCERT 05_모르면 손해보는 Android 디버깅/분석 꿀팁 대방출
200819 NAVER TECH CONCERT 09_Case.xcodeproj - 좋은 동료로 거듭나기 위한 노하우
200820 NAVER TECH CONCERT 14_야 너두 할 수 있어. 비전공자, COBOL 개발자를 거쳐 네이버에서 FE 개발하게 된...
200820 NAVER TECH CONCERT 13_네이버에서 오픈 소스 개발을 통해 성장하는 방법
200820 NAVER TECH CONCERT 12_상반기 네이버 인턴을 돌아보며
200820 NAVER TECH CONCERT 11_빠르게 성장하는 슈퍼루키로 거듭나기
200819 NAVER TECH CONCERT 07_신입 iOS 개발자 개발업무 적응기
Ad

Recently uploaded (20)

PDF
7 ChatGPT Prompts to Help You Define Your Ideal Customer Profile.pdf
PDF
TokAI - TikTok AI Agent : The First AI Application That Analyzes 10,000+ Vira...
PPTX
A Presentation on Artificial Intelligence
PDF
MIND Revenue Release Quarter 2 2025 Press Release
PPTX
Spectroscopy.pptx food analysis technology
PPTX
20250228 LYD VKU AI Blended-Learning.pptx
PDF
Agricultural_Statistics_at_a_Glance_2022_0.pdf
PDF
Optimiser vos workloads AI/ML sur Amazon EC2 et AWS Graviton
PDF
Mobile App Security Testing_ A Comprehensive Guide.pdf
PDF
Reach Out and Touch Someone: Haptics and Empathic Computing
PDF
Blue Purple Modern Animated Computer Science Presentation.pdf.pdf
PDF
Chapter 3 Spatial Domain Image Processing.pdf
DOCX
The AUB Centre for AI in Media Proposal.docx
PPTX
Cloud computing and distributed systems.
PDF
Machine learning based COVID-19 study performance prediction
PDF
NewMind AI Weekly Chronicles - August'25-Week II
PPT
“AI and Expert System Decision Support & Business Intelligence Systems”
PPTX
MYSQL Presentation for SQL database connectivity
PDF
Advanced methodologies resolving dimensionality complications for autism neur...
PDF
Electronic commerce courselecture one. Pdf
7 ChatGPT Prompts to Help You Define Your Ideal Customer Profile.pdf
TokAI - TikTok AI Agent : The First AI Application That Analyzes 10,000+ Vira...
A Presentation on Artificial Intelligence
MIND Revenue Release Quarter 2 2025 Press Release
Spectroscopy.pptx food analysis technology
20250228 LYD VKU AI Blended-Learning.pptx
Agricultural_Statistics_at_a_Glance_2022_0.pdf
Optimiser vos workloads AI/ML sur Amazon EC2 et AWS Graviton
Mobile App Security Testing_ A Comprehensive Guide.pdf
Reach Out and Touch Someone: Haptics and Empathic Computing
Blue Purple Modern Animated Computer Science Presentation.pdf.pdf
Chapter 3 Spatial Domain Image Processing.pdf
The AUB Centre for AI in Media Proposal.docx
Cloud computing and distributed systems.
Machine learning based COVID-19 study performance prediction
NewMind AI Weekly Chronicles - August'25-Week II
“AI and Expert System Decision Support & Business Intelligence Systems”
MYSQL Presentation for SQL database connectivity
Advanced methodologies resolving dimensionality complications for autism neur...
Electronic commerce courselecture one. Pdf

Hierarchical Novelty Detection for Visual Object Recognition

  • 1. Hierarchical Novelty Detection for Visual Object Recognition Kibok Lee*, Kimin Lee†, Kyle Min*, Yuting Zhang*, Jinwoo Shin†, Honglak Lee*‡ University of Michigan*, KAIST†, Google Brain‡
  • 2. Why Hierarchical Novelty Detection? • Conventional novelty detection framework does not provide more information than “novelty” of an object. • Suppose we have training data like … Pomeranian Welsh corgiPersian cat Siamese cat
  • 3. Why Hierarchical Novelty Detection? • Then, suppose we have test images like … Test image: True label: Siamese cat Angora cat Dachshund Pika Prior works: Siamese cat novel novel novel Pomeranian Welsh corgiPersian cat Siamese cat
  • 4. Why Hierarchical Novelty Detection? Ours: Siamese cat novel cat novel dog novel animal animal dogcat Pomeranian Welsh corgiPersian cat Siamese cat True label: Siamese cat Angora cat Dachshund Pika Test image: Prior works: Siamese cat novel novel novel
  • 5. Why Hierarchical Novelty Detection? • Our hierarchical novelty detection framework aims to find the most specific class label of any data on the hierarchical taxonomy built with known labels. Ours: Siamese cat novel cat novel dog novel animal True label: Siamese cat Angora cat Dachshund Pika Test image: Prior works: Siamese cat novel novel novel
  • 6. Why Hierarchical Novelty Detection? • This framework can be potentially useful for automatically or interactively organizing a customized taxonomy • Company’s product catalog • Wildlife monitoring • Personal photo library • by suggesting closest categories for an image from novel categories. • new consumer products • unregistered animal species • untagged scenes or places
  • 7. Hierarchical Taxonomy • Our taxonomy has three types of classes. • Known leaf classes are seen during training • Super classes are ancestors of leaf classes, also known • Novel classes are unseen during training • Their expected prediction is the closest super class in the taxonomy in our task
  • 8. Hierarchical Taxonomy Taxonomy of known classes Set of all known leaf classes Set of parents Set of novel classes whose closest known class is Set of children Set of known classes except and its descendants Set of ancestors including itself:
  • 9. Hierarchical Taxonomy Taxonomy of known classes Set of all known leaf classes Set of parents Set of novel classes whose closest known class is Set of children Set of known classes except and its descendants Set of ancestors including itself:
  • 10. Hierarchical Taxonomy Taxonomy of known classes Set of all known leaf classes Set of parents Set of novel classes whose closest known class is Set of children Set of known classes except and its descendants Set of ancestors including itself:
  • 11. Hierarchical Taxonomy Taxonomy of known classes Set of all known leaf classes Set of parents Set of novel classes whose closest known class is Set of children Set of known classes except and its descendants Set of ancestors including itself:
  • 12. Hierarchical Taxonomy Taxonomy of known classes Set of all known leaf classes Set of parents Set of novel classes whose closest known class is Set of children Set of known classes except and its descendants Set of ancestors including itself:
  • 13. Hierarchical Taxonomy Taxonomy of known classes Set of all known leaf classes Set of parents Set of novel classes whose closest known class is Set of children Set of known classes except and its descendants Set of ancestors including itself:
  • 14. Approach - Top-down (TD) Method • Multi-stage classification • Until arriving at a known leaf class .2 .8 .7 .1 .2
  • 15. Approach - Top-down (TD) Method • Multi-stage classification • Until arriving at a known leaf class • Or classification is unconfident (novel) .5 .5
  • 16. Approach - Top-down (TD) Method • Classification rule: • Classification is confident if
  • 17. Approach - Top-down (TD) Method • Training objective:
  • 18. Approach - Flatten Method • Represent all probabilities of known leaf and novel classes in a single vector • Add virtual novel classes
  • 19. Approach - Flatten Method • Represent all probabilities of known leaf and novel classes in a single vector • Add virtual novel classes • And then flatten the structure
  • 20. Approach - Flatten Method • Represent all probabilities of known leaf and novel classes in a single vector • Add virtual novel classes • And then flatten the structure
  • 21. Approach - Flatten Method • Classification rule: • We propose two strategies to train this model.
  • 22. Approach - Flatten Method • Data relabeling • Fill novel classes by hierarchical relabeling animal cat dog Persian cat Siamese cat Pomeranian Welsh corgi
  • 23. Approach - Flatten Method • Data relabeling • Fill novel classes by hierarchical relabeling animal cat dog Persian cat Siamese cat Pomeranian Welsh corgi
  • 24. Approach - Flatten Method • Data relabeling • Fill novel classes by hierarchical relabeling animal cat dog Persian cat Siamese cat Pomeranian Welsh corgi
  • 25. Approach - Flatten Method • Data relabeling • Fill novel classes by hierarchical relabeling • Relabeling rate can be chosen by validation animal cat dog Persian cat Siamese cat Pomeranian Welsh corgi
  • 26. Approach - Flatten Method • Data relabeling • Training objective: animal cat dog Persian cat Siamese cat Pomeranian Welsh corgi
  • 27. Approach - Flatten Method • Leave-one-out (LOO) strategy • Generate deficient taxonomies • And then train the model with them animal cat dog Persian cat Siamese cat Pomeranian Welsh corgi
  • 28. Approach - Flatten Method • Leave-one-out (LOO) strategy • e.g., when animal cat dog Persian cat Siamese cat Pomeranian Welsh corgi
  • 29. Approach - Flatten Method • Leave-one-out (LOO) strategy • e.g., when animal novel animal dog Pomeranian Welsh corgi
  • 30. Approach - Flatten Method • Leave-one-out (LOO) strategy • Training objective: animal novel animal dog Pomeranian Welsh corgi
  • 31. Approach - Combined Method (TD+LOO) • Compute and enumerate the output of TD • And then feed it to LOO • Combined method utilizes their complementary benefits. • Top-down method leverages hierarchical structure information. • But it suffers from error aggregation over hierarchy. • Flatten method avoids error aggregation over hierarchy. • But it does not leverage hierarchical structure information.
  • 32. Experiments - Hierarchical Novelty Detection • Quantitative results Method ImageNet AwA2 CUB Novel AUC Novel AUC Novel AUC DARTS 10.89 8.83 36.75 35.14 40.42 30.07 Relabel 15.29 11.51 45.71 40.28 38.23 28.75 LOO 15.72 12.00 50.00 43.63 40.78 31.92 TD+LOO 18.78 13.98 53.57 46.77 43.29 33.16
  • 33. Experiments - Hierarchical Novelty Detection • Compared algorithms • Baseline: DARTS (Deng et al., 2012) • Ours: Relabel, LOO, TD+LOO Method ImageNet AwA2 CUB Novel AUC Novel AUC Novel AUC DARTS 10.89 8.83 36.75 35.14 40.42 30.07 Relabel 15.29 11.51 45.71 40.28 38.23 28.75 LOO 15.72 12.00 50.00 43.63 40.78 31.92 TD+LOO 18.78 13.98 53.57 46.77 43.29 33.16 J. Deng, J. Krause, A. C. Berg, and L. Fei-Fei. “Hedging your bets: Optimizing accuracy-specificity trade-offs in large scale visual recognition.” In CVPR, 2012
  • 34. Experiments - Hierarchical Novelty Detection • Datasets • ImageNet: 1k known, 16k novel classes • AwA2: 40 known, 10 novel classes • CUB: 150 known, 50 novel classes Method ImageNet AwA2 CUB Novel AUC Novel AUC Novel AUC DARTS 10.89 8.83 36.75 35.14 40.42 30.07 Relabel 15.29 11.51 45.71 40.28 38.23 28.75 LOO 15.72 12.00 50.00 43.63 40.78 31.92 TD+LOO 18.78 13.98 53.57 46.77 43.29 33.16
  • 35. Experiments - Hierarchical Novelty Detection • Metrics • Novel class accuracy @ known class accuracy = 50% • By adding an appropriate score bias to all novel classes • Area under known-novel class accuracy curve • By varying the novel class score bias Method ImageNet AwA2 CUB Novel AUC Novel AUC Novel AUC DARTS 10.89 8.83 36.75 35.14 40.42 30.07 Relabel 15.29 11.51 45.71 40.28 38.23 28.75 LOO 15.72 12.00 50.00 43.63 40.78 31.92 TD+LOO 18.78 13.98 53.57 46.77 43.29 33.16
  • 36. Experiments - Hierarchical Novelty Detection • Metrics • Novel class accuracy @ known class accuracy = 50% • By adding an appropriate score bias to all novel classes • Area under known-novel class accuracy curve • By varying the novel class score bias (a) ImageNet (b) AwA2 (c) CUB
  • 37. Experiments - Hierarchical Novelty Detection • Metrics • Novel class accuracy @ known class accuracy = 50% • By adding an appropriate score bias to all novel classes • Area under known-novel class accuracy curve • By varying the novel class score bias (a) ImageNet (b) AwA2 (c) CUB
  • 38. Experiments - Hierarchical Novelty Detection • Metrics • Novel class accuracy @ known class accuracy = 50% • By adding an appropriate score bias to all novel classes • Area under known-novel class accuracy curve • By varying the novel class score bias (a) ImageNet (b) AwA2 (c) CUB
  • 39. Experiments - Hierarchical Novelty Detection • Qualitative results Novel class: American foxhound GT foxhound
  • 40. Experiments - Hierarchical Novelty Detection • Qualitative results Novel class: American foxhound Method ε A Word GT foxhound DARTS 2 N beagle Relabel 1 Y hound dog LOO 0 Y foxhound TD+LOO 0 Y foxhound
  • 41. Experiments - Hierarchical Novelty Detection • Qualitative results Novel class: serval Method ε A Word GT wildcat DARTS 3 N Egyptian cat Relabel 2 N domestic cat LOO 2 Y feline TD+LOO 1 Y cat
  • 42. Experiments - Hierarchical Novelty Detection • Qualitative results Novel class: song thrush Method ε A Word GT thrush DARTS 3 N hummingbird Relabel 2 Y bird LOO 1 Y oscine bird TD+LOO 0 Y thrush
  • 43. Experiments - Hierarchical Novelty Detection • Qualitative results Novel class: ice-cream sundae Method ε A Word GT frozen dessert DARTS 4 Y food, nutrient Relabel 1 N ice cream LOO 1 Y dessert TD+LOO 0 Y frozen dessert
  • 44. Experiments - Generalized Zero-Shot Learning • Quantitative results Embedding AwA1 AwA2 CUB Att Word Hier Unseen AUC Unseen AUC Unseen AUC √ 65.29 50.02 63.87 51.27 50.05 23.60 √ 51.87 39.67 54.77 42.21 27.28 11.47 √ √ 67.80 52.84 65.76 53.18 49.83 24.13 Path 42.57 30.58 44.34 33.44 24.22 8.38 √ Path 67.09 51.45 66.58 53.50 50.25 23.70 √ Path 52.89 40.66 55.28 42.86 27.72 11.65 √ √ Path 68.04 53.21 67.28 54.31 50.87 24.20 TD 33.86 25.56 31.84 24.97 13.09 7.20 √ TD 66.13 54.66 66.86 57.49 50.17 30.31 √ TD 56.14 46.28 59.67 49.39 29.05 16.73 √ √ TD 69.23 57.67 68.80 59.24 50.17 30.31
  • 45. Experiments - Generalized Zero-Shot Learning • Semantic embeddings • Attributes (numeric attribute values) • Word vector (similarity among words in real coordinate space) • Hierarchical embedding Embedding AwA1 AwA2 CUB Att Word Hier Unseen AUC Unseen AUC Unseen AUC √ 65.29 50.02 63.87 51.27 50.05 23.60 √ 51.87 39.67 54.77 42.21 27.28 11.47 √ √ 67.80 52.84 65.76 53.18 49.83 24.13 Path 42.57 30.58 44.34 33.44 24.22 8.38 √ Path 67.09 51.45 66.58 53.50 50.25 23.70 √ Path 52.89 40.66 55.28 42.86 27.72 11.65 √ √ Path 68.04 53.21 67.28 54.31 50.87 24.20 TD 33.86 25.56 31.84 24.97 13.09 7.20 √ TD 66.13 54.66 66.86 57.49 50.17 30.31 √ TD 56.14 46.28 59.67 49.39 29.05 16.73 √ √ TD 69.23 57.67 68.80 59.24 50.17 30.31
  • 46. Experiments - Generalized Zero-Shot Learning • Compared hierarchical embeddings • Baseline: Path (Akata et al., 2015) • Distance between classes on hierarchy • Ours: Top-down (TD) • Expected output of our top-down model Embedding AwA1 AwA2 CUB Att Word Hier Unseen AUC Unseen AUC Unseen AUC √ 65.29 50.02 63.87 51.27 50.05 23.60 √ 51.87 39.67 54.77 42.21 27.28 11.47 √ √ 67.80 52.84 65.76 53.18 49.83 24.13 Path 42.57 30.58 44.34 33.44 24.22 8.38 √ Path 67.09 51.45 66.58 53.50 50.25 23.70 √ Path 52.89 40.66 55.28 42.86 27.72 11.65 √ √ Path 68.04 53.21 67.28 54.31 50.87 24.20 TD 33.86 25.56 31.84 24.97 13.09 7.20 √ TD 66.13 54.66 66.86 57.49 50.17 30.31 √ TD 56.14 46.28 59.67 49.39 29.05 16.73 √ √ TD 69.23 57.67 68.80 59.24 50.17 30.31 Z. Akata, S. Reed, D. Walter, H. Lee, and B. Schiele. “Evaluation of output embeddings for fine-grained image classification.” In CVPR, 2015.
  • 47. Experiments - Generalized Zero-Shot Learning • Datasets • AwA1,2: 40 known, 10 novel classes • CUB: 150 known, 50 novel classes Embedding AwA1 AwA2 CUB Att Word Hier Unseen AUC Unseen AUC Unseen AUC √ 65.29 50.02 63.87 51.27 50.05 23.60 √ 51.87 39.67 54.77 42.21 27.28 11.47 √ √ 67.80 52.84 65.76 53.18 49.83 24.13 Path 42.57 30.58 44.34 33.44 24.22 8.38 √ Path 67.09 51.45 66.58 53.50 50.25 23.70 √ Path 52.89 40.66 55.28 42.86 27.72 11.65 √ √ Path 68.04 53.21 67.28 54.31 50.87 24.20 TD 33.86 25.56 31.84 24.97 13.09 7.20 √ TD 66.13 54.66 66.86 57.49 50.17 30.31 √ TD 56.14 46.28 59.67 49.39 29.05 16.73 √ √ TD 69.23 57.67 68.80 59.24 50.17 30.31
  • 48. Experiments - Generalized Zero-Shot Learning • Metrics • Unseen class accuracy (ZSL) • Area under seen-unseen curve (GZSL) • By varying the unseen class score bias Embedding AwA1 AwA2 CUB Att Word Hier Unseen AUC Unseen AUC Unseen AUC √ 65.29 50.02 63.87 51.27 50.05 23.60 √ 51.87 39.67 54.77 42.21 27.28 11.47 √ √ 67.80 52.84 65.76 53.18 49.83 24.13 Path 42.57 30.58 44.34 33.44 24.22 8.38 √ Path 67.09 51.45 66.58 53.50 50.25 23.70 √ Path 52.89 40.66 55.28 42.86 27.72 11.65 √ √ Path 68.04 53.21 67.28 54.31 50.87 24.20 TD 33.86 25.56 31.84 24.97 13.09 7.20 √ TD 66.13 54.66 66.86 57.49 50.17 30.31 √ TD 56.14 46.28 59.67 49.39 29.05 16.73 √ √ TD 69.23 57.67 68.80 59.24 50.17 30.31
  • 49. Experiments - Generalized Zero-Shot Learning • Metrics • Unseen class accuracy (ZSL) • Area under seen-unseen curve (GZSL) • By varying the unseen class score bias (a) AwA1 (b) AwA2 (c) CUB
  • 50. Experiments - Generalized Zero-Shot Learning • Metrics • Unseen class accuracy (ZSL) • Area under seen-unseen curve (GZSL) • By varying the unseen class score bias (a) AwA1 (b) AwA2 (c) CUB
  • 51. Experiments - Generalized Zero-Shot Learning • Metrics • Unseen class accuracy (ZSL) • Area under seen-unseen curve (GZSL) • By varying the unseen class score bias (a) AwA1 (b) AwA2 (c) CUB
  • 52. Conclusion • We propose a hierarchical novelty detection framework, which aims to find the most specific class label of any data on the hierarchical taxonomy built with known labels. • We propose two approaches. • Top-down method performs classification & novelty detection hierarchically. • Flatten method computes single probability vector of all candidate classes. • Their combination takes advantage of their complementary benefits. • Our model can be combined with other semantic embeddings to improve generalized zero-shot learning performance.