SlideShare a Scribd company logo
GRAPH REGULARISED HASHING
SEAN MORAN†, VICTOR LAVRENKO
† SEAN.MORAN@ED.AC.UK
RESEARCH QUESTION
• Locality sensitive hashing (LSH) [1] fractures the input feature
space space with randomly placed hyperplanes.
• Can we do better by using supervision to adjust the hyperplanes?
INTRODUCTION
• Problem: Constant time nearest-neighbour search in large datasets.
• Hashing-based approximate nearest neighbour (NN) search:
– Index image/documents into the buckets of hashtable(s)
– Encourage collisions between similar images/documents
110101
010111
111101
H
H
Content Based IR
Image: Imense Ltd
Image: Doersch et al.
Image: Xu et al.
Location Recognition
Near duplicate detection
010101
111101
.....
H
QUERY
DATABASE
QUERY
NEAREST
NEIGHBOURS
HASH TABLE
COMPUTE
SIMILARITY
• Advantages:
– O(1) lookup per query rather than O(N) (brute-force)
– Memory/storage saving due to compact binary codes
GRAPH REGULARISED HASHING (GRH)
• We propose a two step iterative hashing model, Graph Regularised
Hashing (GRH) [5]. GRH uses supervision in the form of an adja-
cency matrix that specifies whether or not data-points are related.
– Step A: Graph Regularisation: the K-bit hashcode of a data-
point is set to the average of the data-points of its nearest
neighbours as specified by the adjacency graph:
Lm ← sgn α SD−1
Lm−1 + (1−α)L0
∗ S: Affinity (adjacency) matrix
∗ D: Diagonal degree matrix
∗ L: Binary bits at iteration m
∗ α ∈ {0, 1}: Linear interpolation parameter
– Step A is a simple sparse-sparse matrix multiplication, and can
be implemented very efficiently. Any existing hash function
e.g. LSH [1] can be used to initialise the bits in L0
– Step B: Data-Space Partitioning: the hashcodes produced in
Step A are used as the labels to learn K binary classifiers. This
is the out-of-sample extension step, allowing the encoding of
data-points not seen before:
for k = 1. . .K : min ||wk||2
+ C
Ntrd
i=1 ξik
s.t. Lik(wk xi + tk) ≥ 1 − ξik for i = 1. . .Ntrd
∗ wk: Hyperplane k tk: bias k
∗ xi: data-point i Lik: bit k of data-point i
∗ ξik: slack variable K: # bits Ntrd: # data-points
• Steps A-B are repeated for a set number of iterations (M) (e.g. < 10).
The learnt hyperplanes wk can then be used to encode unseen data-
points (via a simple dot-product).
STEP A: GRAPH REGULARISATION
• Toy example: nodes are images with 3-bit LSH encoding. Arcs
indicate nearest neighbour relationships. We show two images
(c,e) having their hashcodes updated in Step A:
-1 1 1
1 1 1
-1 -1 -1
1 1 -1
ba
c
e
f
g
h
d
-1 1 1
1 -1 -1
1 -1 -1
1 1 1 -1 1 1
1 1 -1
STEP B: DATA-SPACE PARTITIONING
• Here we show a hyperplane being learnt using the first bit as
(highlighted with bold box) as label. One hyperplane is learnt
per bit.
-1 1 1
1 1 1
-1 -1 -1
1 1 -1
ba
c
e
f
g
h
d
-1 1 1
w1. x−t1=0
w1
Negative
(-1)
half space
Positive
(+1)
half space
1 1 -1
1 -1 -1
-1 1 1
QUANTITATIVE RESULTS (CIFAR-10) (MORE DATASETS IN PAPER)
• Mean average precision (mAP) image retrieval results using
GIST features on CIFAR-10 ( : is significant at p < 0.01):
CIFAR-10
16 bits 32 bits 48 bits 64 bits
ITQ+CCA [2] 0.2015 0.2130 0.2208 0.2237
STH [3] 0.2352 0.2072 0.2118 0.2000
KSH [4] 0.2496 0.2785 0.2849 0.2905
GRH [5] 0.2991 0.3122 0.3252 0.3350
• Timings (seconds) averaged over 10 runs. GRH is 1) faster to
train and 2) is faster to encode unseen data-points:
Training Testing Total
GRH [5] 8.01 0.03 8.04
KSH [4] 74.02 0.10 74.12
BRE [6] 227.84 0.37 228.21
SUMMARY OF KEY FINDINGS
• First both accurate and scalable supervised hashing model
• Future work will extend GRH to streaming data sources
• Code online: https://github.com/sjmoran/grh
References:
[1] P. Indyk, R. Motwani: Approximate nearest neighbors: Towards removing the curse of dimensionality. In: STOC (1998).
[2] Y. Gong, S. Lazebnik: Iterative Quantisation. In: CVPR (2011). , [3] D. Zhang et al. Self-Taught Hashing. In: SIGIR (2010)., [4] W. Liu et
al. Supervised Hashing with Kernels. In: CVPR (2012)., [5] S. Moran, V. Lavrenko. Graph Regularised Hashing. In: ECIR (2015).,
[6] B. Kulis, T. Darrell et al. Binary Reconstructive Embedding. In: NIPS (2009).

More Related Content

PDF
Learning to Project and Binarise for Hashing-based Approximate Nearest Neighb...
PPTX
IOEfficientParalleMatrixMultiplication_present
PPT
PPTX
K10692 control theory
PPTX
Digit recognizer by convolutional neural network
PPT
Towards Utilizing GPUs in Information Visualization
PPTX
Big spatial2014 mapreduceweights
PDF
Joint CSI Estimation, Beamforming and Scheduling Design for Wideband Massive ...
Learning to Project and Binarise for Hashing-based Approximate Nearest Neighb...
IOEfficientParalleMatrixMultiplication_present
K10692 control theory
Digit recognizer by convolutional neural network
Towards Utilizing GPUs in Information Visualization
Big spatial2014 mapreduceweights
Joint CSI Estimation, Beamforming and Scheduling Design for Wideband Massive ...

What's hot (20)

PDF
Neighbourhood Preserving Quantisation for LSH SIGIR Poster
PPTX
[Vldb 2013] skyline operator on anti correlated distributions
PPTX
distance_matrix_ch
PPTX
PCL (Point Cloud Library)
PDF
Andrew Goldberg. Highway Dimension and Provably Efficient Shortest Path Algor...
PDF
ES_SAA_OG_PF_ECCTD_Pos
ODP
Deep Learning meetup
ODP
Hubba Deep Learning
PPT
Mmclass5
PDF
Faster Practical Block Compression for Rank/Select Dictionaries
PDF
Deep single view 3 d object reconstruction with visual hull
PDF
Quiz 2
PPTX
Md2k 0219 shang
PPTX
Point cloud library
PDF
Nonlinear dimension reduction
PDF
Discovering human places of interest from multimodal mobile phone data
TXT
Saga.lng
PDF
(Paper Review)3D shape reconstruction from sketches via multi view convolutio...
PDF
06.09.2017 Computer Science, Machine Learning & Statistiks Meetup - MULTI-GPU...
Neighbourhood Preserving Quantisation for LSH SIGIR Poster
[Vldb 2013] skyline operator on anti correlated distributions
distance_matrix_ch
PCL (Point Cloud Library)
Andrew Goldberg. Highway Dimension and Provably Efficient Shortest Path Algor...
ES_SAA_OG_PF_ECCTD_Pos
Deep Learning meetup
Hubba Deep Learning
Mmclass5
Faster Practical Block Compression for Rank/Select Dictionaries
Deep single view 3 d object reconstruction with visual hull
Quiz 2
Md2k 0219 shang
Point cloud library
Nonlinear dimension reduction
Discovering human places of interest from multimodal mobile phone data
Saga.lng
(Paper Review)3D shape reconstruction from sketches via multi view convolutio...
06.09.2017 Computer Science, Machine Learning & Statistiks Meetup - MULTI-GPU...
Ad

Viewers also liked (20)

PPT
Digital Techniques
PPTX
Strings & arrays
PDF
Learn to Write ur first Shell script
PPT
Introduction to c_language
PDF
Storage devices
PDF
Pemrograman shell2
PPTX
ppt on power supplies by prince kumar kusshwaha(RJIT)
PPTX
A brief introduction to C Language
PPTX
Adc &amp;dac ppt
PDF
Data Structure (Introduction to Data Structure)
PDF
CSEC Physics Review - Introduction To Logic Gates
PPTX
Direct linking loader
PPTX
System programming
PPT
Learning sed and awk
PDF
Introduction to data structure
PPTX
6. Linked list - Data Structures using C++ by Varsha Patil
PPT
Basic 50 linus command
PDF
Compiler Design Introduction
PPTX
Pointer in c program
PPTX
Digital Techniques
Strings & arrays
Learn to Write ur first Shell script
Introduction to c_language
Storage devices
Pemrograman shell2
ppt on power supplies by prince kumar kusshwaha(RJIT)
A brief introduction to C Language
Adc &amp;dac ppt
Data Structure (Introduction to Data Structure)
CSEC Physics Review - Introduction To Logic Gates
Direct linking loader
System programming
Learning sed and awk
Introduction to data structure
6. Linked list - Data Structures using C++ by Varsha Patil
Basic 50 linus command
Compiler Design Introduction
Pointer in c program
Ad

Similar to Graph Regularised Hashing (20)

PPTX
ImageNet classification with deep convolutional neural networks(2012)
PDF
Regularised Cross-Modal Hashing (SIGIR'15 Poster)
PDF
Graph Regularised Hashing (ECIR'15 Talk)
PDF
NVIDIA 深度學習教育機構 (DLI): Image segmentation with tensorflow
PDF
G-Store: High-Performance Graph Store for Trillion-Edge Processing
PDF
Tutorial-on-DNN-07-Co-design-Precision.pdf
PPTX
IEEE_VTC_Bhaskar_LiDAR_Data_Compressoionpptx
PPTX
lossy compression JPEG
PDF
Hardware Acceleration for Machine Learning
PDF
superglue_slides.pdf
PDF
Practical Spherical Harmonics Based PRT Methods
PPTX
Introduction to computer vision
PDF
Enterprise Scale Topological Data Analysis Using Spark
PDF
Enterprise Scale Topological Data Analysis Using Spark
PPT
lecture3.pptlecture3 data structures pptt
PPTX
Introduction to computer vision with Convoluted Neural Networks
PDF
Scratch to Supercomputers: Bottoms-up Build of Large-scale Computational Lens...
PDF
Mask-RCNN for Instance Segmentation
PDF
"Fundamentals of Monocular SLAM," a Presentation from Cadence
PDF
Digital Image Procesing
ImageNet classification with deep convolutional neural networks(2012)
Regularised Cross-Modal Hashing (SIGIR'15 Poster)
Graph Regularised Hashing (ECIR'15 Talk)
NVIDIA 深度學習教育機構 (DLI): Image segmentation with tensorflow
G-Store: High-Performance Graph Store for Trillion-Edge Processing
Tutorial-on-DNN-07-Co-design-Precision.pdf
IEEE_VTC_Bhaskar_LiDAR_Data_Compressoionpptx
lossy compression JPEG
Hardware Acceleration for Machine Learning
superglue_slides.pdf
Practical Spherical Harmonics Based PRT Methods
Introduction to computer vision
Enterprise Scale Topological Data Analysis Using Spark
Enterprise Scale Topological Data Analysis Using Spark
lecture3.pptlecture3 data structures pptt
Introduction to computer vision with Convoluted Neural Networks
Scratch to Supercomputers: Bottoms-up Build of Large-scale Computational Lens...
Mask-RCNN for Instance Segmentation
"Fundamentals of Monocular SLAM," a Presentation from Cadence
Digital Image Procesing

More from Sean Moran (7)

PDF
Deep Local Parametric Filters for Image Enhancement
PDF
Deep Local Parametric Filters for Image Enhancement
PDF
PhD thesis defence slides
PDF
Data Science Research Day (Talk)
PDF
Sparse Kernel Learning for Image Annotation
PDF
ICMR 2014 - Sparse Kernel Learning Poster
PDF
ACL Variable Bit Quantisation Talk
Deep Local Parametric Filters for Image Enhancement
Deep Local Parametric Filters for Image Enhancement
PhD thesis defence slides
Data Science Research Day (Talk)
Sparse Kernel Learning for Image Annotation
ICMR 2014 - Sparse Kernel Learning Poster
ACL Variable Bit Quantisation Talk

Recently uploaded (20)

PDF
Recruitment and Placement PPT.pdfbjfibjdfbjfobj
PDF
Lecture1 pattern recognition............
PPT
Chapter 2 METAL FORMINGhhhhhhhjjjjmmmmmmmmm
PPTX
oil_refinery_comprehensive_20250804084928 (1).pptx
PPT
Quality review (1)_presentation of this 21
PDF
168300704-gasification-ppt.pdfhghhhsjsjhsuxush
PPTX
iec ppt-1 pptx icmr ppt on rehabilitation.pptx
PPT
Chapter 3 METAL JOINING.pptnnnnnnnnnnnnn
PDF
Foundation of Data Science unit number two notes
PPTX
CEE 2 REPORT G7.pptxbdbshjdgsgjgsjfiuhsd
PPTX
Introduction to Firewall Analytics - Interfirewall and Transfirewall.pptx
PDF
TRAFFIC-MANAGEMENT-AND-ACCIDENT-INVESTIGATION-WITH-DRIVING-PDF-FILE.pdf
PPTX
The THESIS FINAL-DEFENSE-PRESENTATION.pptx
PPTX
Business Ppt On Nestle.pptx huunnnhhgfvu
PPT
Reliability_Chapter_ presentation 1221.5784
PPTX
Introduction-to-Cloud-ComputingFinal.pptx
PPTX
Database Infoormation System (DBIS).pptx
PPTX
advance b rammar.pptxfdgdfgdfsgdfgsdgfdfgdfgsdfgdfgdfg
PDF
Fluorescence-microscope_Botany_detailed content
PPTX
ALIMENTARY AND BILIARY CONDITIONS 3-1.pptx
Recruitment and Placement PPT.pdfbjfibjdfbjfobj
Lecture1 pattern recognition............
Chapter 2 METAL FORMINGhhhhhhhjjjjmmmmmmmmm
oil_refinery_comprehensive_20250804084928 (1).pptx
Quality review (1)_presentation of this 21
168300704-gasification-ppt.pdfhghhhsjsjhsuxush
iec ppt-1 pptx icmr ppt on rehabilitation.pptx
Chapter 3 METAL JOINING.pptnnnnnnnnnnnnn
Foundation of Data Science unit number two notes
CEE 2 REPORT G7.pptxbdbshjdgsgjgsjfiuhsd
Introduction to Firewall Analytics - Interfirewall and Transfirewall.pptx
TRAFFIC-MANAGEMENT-AND-ACCIDENT-INVESTIGATION-WITH-DRIVING-PDF-FILE.pdf
The THESIS FINAL-DEFENSE-PRESENTATION.pptx
Business Ppt On Nestle.pptx huunnnhhgfvu
Reliability_Chapter_ presentation 1221.5784
Introduction-to-Cloud-ComputingFinal.pptx
Database Infoormation System (DBIS).pptx
advance b rammar.pptxfdgdfgdfsgdfgsdgfdfgdfgsdfgdfgdfg
Fluorescence-microscope_Botany_detailed content
ALIMENTARY AND BILIARY CONDITIONS 3-1.pptx

Graph Regularised Hashing

  • 1. GRAPH REGULARISED HASHING SEAN MORAN†, VICTOR LAVRENKO † SEAN.MORAN@ED.AC.UK RESEARCH QUESTION • Locality sensitive hashing (LSH) [1] fractures the input feature space space with randomly placed hyperplanes. • Can we do better by using supervision to adjust the hyperplanes? INTRODUCTION • Problem: Constant time nearest-neighbour search in large datasets. • Hashing-based approximate nearest neighbour (NN) search: – Index image/documents into the buckets of hashtable(s) – Encourage collisions between similar images/documents 110101 010111 111101 H H Content Based IR Image: Imense Ltd Image: Doersch et al. Image: Xu et al. Location Recognition Near duplicate detection 010101 111101 ..... H QUERY DATABASE QUERY NEAREST NEIGHBOURS HASH TABLE COMPUTE SIMILARITY • Advantages: – O(1) lookup per query rather than O(N) (brute-force) – Memory/storage saving due to compact binary codes GRAPH REGULARISED HASHING (GRH) • We propose a two step iterative hashing model, Graph Regularised Hashing (GRH) [5]. GRH uses supervision in the form of an adja- cency matrix that specifies whether or not data-points are related. – Step A: Graph Regularisation: the K-bit hashcode of a data- point is set to the average of the data-points of its nearest neighbours as specified by the adjacency graph: Lm ← sgn α SD−1 Lm−1 + (1−α)L0 ∗ S: Affinity (adjacency) matrix ∗ D: Diagonal degree matrix ∗ L: Binary bits at iteration m ∗ α ∈ {0, 1}: Linear interpolation parameter – Step A is a simple sparse-sparse matrix multiplication, and can be implemented very efficiently. Any existing hash function e.g. LSH [1] can be used to initialise the bits in L0 – Step B: Data-Space Partitioning: the hashcodes produced in Step A are used as the labels to learn K binary classifiers. This is the out-of-sample extension step, allowing the encoding of data-points not seen before: for k = 1. . .K : min ||wk||2 + C Ntrd i=1 ξik s.t. Lik(wk xi + tk) ≥ 1 − ξik for i = 1. . .Ntrd ∗ wk: Hyperplane k tk: bias k ∗ xi: data-point i Lik: bit k of data-point i ∗ ξik: slack variable K: # bits Ntrd: # data-points • Steps A-B are repeated for a set number of iterations (M) (e.g. < 10). The learnt hyperplanes wk can then be used to encode unseen data- points (via a simple dot-product). STEP A: GRAPH REGULARISATION • Toy example: nodes are images with 3-bit LSH encoding. Arcs indicate nearest neighbour relationships. We show two images (c,e) having their hashcodes updated in Step A: -1 1 1 1 1 1 -1 -1 -1 1 1 -1 ba c e f g h d -1 1 1 1 -1 -1 1 -1 -1 1 1 1 -1 1 1 1 1 -1 STEP B: DATA-SPACE PARTITIONING • Here we show a hyperplane being learnt using the first bit as (highlighted with bold box) as label. One hyperplane is learnt per bit. -1 1 1 1 1 1 -1 -1 -1 1 1 -1 ba c e f g h d -1 1 1 w1. x−t1=0 w1 Negative (-1) half space Positive (+1) half space 1 1 -1 1 -1 -1 -1 1 1 QUANTITATIVE RESULTS (CIFAR-10) (MORE DATASETS IN PAPER) • Mean average precision (mAP) image retrieval results using GIST features on CIFAR-10 ( : is significant at p < 0.01): CIFAR-10 16 bits 32 bits 48 bits 64 bits ITQ+CCA [2] 0.2015 0.2130 0.2208 0.2237 STH [3] 0.2352 0.2072 0.2118 0.2000 KSH [4] 0.2496 0.2785 0.2849 0.2905 GRH [5] 0.2991 0.3122 0.3252 0.3350 • Timings (seconds) averaged over 10 runs. GRH is 1) faster to train and 2) is faster to encode unseen data-points: Training Testing Total GRH [5] 8.01 0.03 8.04 KSH [4] 74.02 0.10 74.12 BRE [6] 227.84 0.37 228.21 SUMMARY OF KEY FINDINGS • First both accurate and scalable supervised hashing model • Future work will extend GRH to streaming data sources • Code online: https://github.com/sjmoran/grh References: [1] P. Indyk, R. Motwani: Approximate nearest neighbors: Towards removing the curse of dimensionality. In: STOC (1998). [2] Y. Gong, S. Lazebnik: Iterative Quantisation. In: CVPR (2011). , [3] D. Zhang et al. Self-Taught Hashing. In: SIGIR (2010)., [4] W. Liu et al. Supervised Hashing with Kernels. In: CVPR (2012)., [5] S. Moran, V. Lavrenko. Graph Regularised Hashing. In: ECIR (2015)., [6] B. Kulis, T. Darrell et al. Binary Reconstructive Embedding. In: NIPS (2009).