SlideShare a Scribd company logo
Fast matrix primitives for
ranking, communities !
and more.



David F. Gleich!

David Gleich · Purdue

Netflix

1

Computer Science!
Purdue University!
Netflix

2

David Gleich · Purdue
error
1

Models Previous work
– and algorithms for high performance !
from the PI tackled net- computations
matrix and network
FIGURE 6

std

2

work alignment with matrix methods =for cm
edge
Std, s 0.39
Big data(b)methods
overlap:
SIMAX ‘09, SISC ‘11,MapReduce ‘11, ICASSP ’12
1

j

i

i0
Overlap
Overlap

j0

error



SC ‘05, WAW ‘07, SISC ‘10, WWW ’10, …

Massive matrix "
computations

std

0

0

Fast & Scalable"
Network centrality

0

20
10

10

A
L
B
Tensor eigenvalues"
0

(d) Std, s = 1.95 cm

Ax = b
min kAx bk
Ax = x

This proposal is for matchand a power method
Network alignment
tensor
ing triangles using

P
methods:
on multi-threaded
maximize
Tijk xi xj xk

model compared to the prediction standard debble locations at the final time for two values of
ICDM ‘09, SC ‘11, TKDE ‘13
= 1.95 cm. (Colors are visible in the electronic

ijk
n
and kxk2 = 1
subject to distributed 
j
Triangle
j
X
i
s
i
approximately twenty minutes to construct using (next) architectures
k
[x
]i = ⇢ · (
Tijk xj xk + xi )
k
s.
jk
s- involved a few pre- and post-processing steps:
ta
where ! ensures the 2-norm
m Aria, globally transpose the data, compute the
g errors. The preprocessing steps took approx- SSHOPM method due to "
nd
0

0

Data clustering

WSDM ‘12, KDD ‘12, CIKM ’13 …

0

A
recise timing information, L we do notB
but
report

David
Kolda and Mayo
Gleich

· Purdue

Netflix

3

t
r
o
s.
g
n.
o
The talk ends, you
believe -- whatever
you want to. 

Image from rockysprings, deviantart, CC share-alike

4

Everything in the world can be
explained by a matrix, and we see
how deep the rabbit hole goes
Matrix computations in a red-pill

David Gleich · Purdue

Netflix

5

Solve a problem better by
exploiting its structure!
Problem 1 – (Faster) !
Recommendation as link prediction
WHY NO PREPROCESSING?
Top-k predicted “links”
are movies to watch!

David F. Gleich (Purdue)

David Gleich · Purdue

Emory Math/CS Seminar

Netflix

6

Pairwise scores give
user similarity

19 of 47
David Gleich · Purdue

Netflix

7

Problem 2 – (Better) !
Best movies
Matrix computations in a red-pill

David Gleich · Purdue

Netflix

8

Solve a problem better by
exploiting its structure!
Matrix structure
Netflix graph

Movies "
“liked”
(>3 stars?)
Problem 1!
Adjacency matrix
Normalized Laplacian matrix
Random walk matrix

Netflix matrix
1

1

4
5

5

Problem 2"
Pairwise comparison matrix
David Gleich · Purdue

Netflix

9

5
Problem 1 – (Faster) !
Recommendation as link prediction
WHY NO PREPROCESSING?
Top-k predicted “links”
are movies to watch!

David F. Gleich (Purdue)

David Gleich · Purdue

Emory Math/CS Seminar

Netflix

10

Pairwise scores give
user similarity

19 of 47
z score (edge-based) is

  movie  =   X  ↵   ✓            
  
pred. on
`

`=1

1
X

num. paths of length `
from user to movie

✓

◆

user
k=
(↵ A )
ind. vec.
`=1
{z
}
|
` Math/CS Seminar
Emory `

⌘ei

◆

David Gleich · Purdue

Netflix

11

1

Movie prediction"
vector

)

Matrix based link predictors
Matrix based link predictors
1
X

✓

◆

user
k=
(↵ A )
ind. vec.
`=1
{z
}
|
`

`

⌘ei

Neumann

Carl Neumann

1
X
k =0

↵A)k = ei

(tA)k
David Gleich · Purdue

Netflix

12

(I
The Katz score (edge-based) is

Matrix based link predictors

                             

(I

↵A)k = ei

PageRank

(I

↵P)x = ei

Semi-super."
learning
Heat kernel

(IDavid F. Gleich (Purdue) = ei
↵L)x
exp{↵P}x = ei

Emory Math/CS Seminar

They all look at sums of "
damped paths, but "
change the details, slightly

David Gleich · Purdue

Netflix

13

Katz
Matrix based link predictors
are localized!
PageRank scores for one node!
Crawl of flickr from 2006 ~800k nodes, 6M edges, alpha=1/2
0

1.5

error
||xtrue – xnnz||1

10

1

0.5

−10

10

−15

0

2

4

plot(x)

6

8

10
5

x 10

10

0

10

2

4

10

6

10

10

nonzeros

David Gleich · Purdue

Netflix

14

0

−5

10
Matrix based link predictors
are localized!
KATZ SCORES ARE LOCALIZED

David F. Gleich (Purdue)

Emory Math/CS Seminar

32 of 47

David Gleich · Purdue

Netflix

15

Up to 50 neighbors is
99.65% of the total
mass
Matrix computations in a red-pill

David Gleich · Purdue

Netflix

16

Solve a problem better by
exploiting its structure!
How do we compute them fast?
PageRank

xj = ↵

X

i neigh. of j

xi
deg(i)

+ 1 if j is the target user

w/ access to in-links & degs.

w/ access to out-links

PageRankPull

PageRankPush

xj(k+1)

(k)
↵xa /6

(k)
↵xb /2

(k)
↵xc /3

= fj
xj(k+1)

↵

X
i!j

xi(k ) /degi = fj

Let 

b
a

c

j = blue node

(k+1)

= xj(k) + rj

(k +1)

=0

then
 xj

Update 
r(k +1) rj

(k
(k)
ra +1) = ra + ↵rj(k ) /3

(k
(k)
rb +1) = rb + ↵rj(k ) /3
(k
(k)
rc +1) = rc + ↵rj(k ) /3

David Gleich · Purdue

Netflix

17

(k +1)
Solve for 
xj

j = blue node
We have good theory
for this algorithm …

David Gleich · Purdue

Netflix

18

… and even better
empirical performance.
Theory
Andersen, Chung, Lang (2006)!
For PageRank, “fast runtimes” and “localization”
Bonchi, Esfandiar, Gleich, et al. (2010/2013)!
For Katz, “fast runtimes” 

David Gleich · Purdue

Netflix

19

Kloster, Gleich (2013)!
For Katz, Heat Kernel, "
“fast runtimes” and “localization”"
(assuming power-law degrees)
Accuracy vs. work !
(Heat kernel)
dblp−cc
dblp collaboration graph, 225k vertices
1

0.6

tol=10−5

tol=10−4

0.4

@10
@25

0.2

@100
@1000

0
−2

−1

0

10
10
10
Effective matrix−vector products

David Gleich · Purdue

Netflix

20

Precision

0.8

For the dblp collaboration
graph, we study the
precision in finding the
100 largest nodes as we
vary the work. This set of
100 does not include the
nodes immediate
neighbors. (One column,
but representative)
David Gleich · Purdue

Netflix

21

Empirical runtime (Katz)
TIMING
40

Never got to try it …
 analytics
test on 
HelloMovies.com Need to ix now
matr
Netflix

60

80

Ran out of money once we had the algorithms
… promising initial results though!

I collaborate with the company behind He
David Gleich · Purdue

Netflix

22

Note

1
1
1
1
1
1
1
1
David Gleich · Purdue

Netflix

23

Problem 2 – (Better) !
Best movies
Which is a better list of good DVDs?
Lord of the Rings 3: The Return of …

Lord of the Rings 3: The Return of …

Lord of the Rings 1: The Fellowship 

Lord of the Rings 1: The Fellowship 

Lord of the Rings 2: The Two Towers

Lord of the Rings 2: The Two Towers

Lost: Season 1

Star Wars V: Empire Strikes Back

Battlestar Galactica: Season 1

Raiders of the Lost Ark

Fullmetal Alchemist

Star Wars IV: A New Hope

Trailer Park Boys: Season 4

Shawshank Redemption

Trailer Park Boys: Season 3

Star Wars VI: Return of the Jedi

Tenchi Muyo!

Lord of the Rings 3: Bonus DVD

Shawshank Redemption

The Godfather
Nuclear Norm "
based rank aggregation

(the mean rating)

(not matrix completion on the
netflix rating matrix)
David Gleich · Purdue

Netflix

24/40

Standard "
rank aggregation"
Rank Aggregation

Given partial orders on subsets of items, rank aggregation
is the problem of finding an overall ordering.

Voting Find the winning candidate

Program committees Find the best papers given reviews
Dining Find the best restaurant in Chicago
David Gleich · Purdue

Netflix

25/40
Ranking is really hard
John Kemeny
Ken Arrow

All rank aggregations
involve some measure of
compromise

A good ranking is the
“average” ranking under a
permutation distance

NP hard to compute
Kemeny’s ranking

David Gleich · Purdue

Netflix

26/40

Dwork, Kumar, Naor, !
Sivikumar
Supposewe had scores
Suppose we had scores
Let    be the score of the ith movie/song/paper/team to rank
Suppose we can compare the ith to jth:

  
is skew-symmetric, rank 2.

Also works for   

with an extra log.

Numerical ranking is intimately intertwined
with skew-symmetric matrices
Kemeny and Snell, Mathematical Models in Social Sciences (1978)
David F. Gleich (Purdue)

KDD 2011

David Gleich · Purdue

Netflix
6/20

27/40

Then   
Using ratings as comparisons

Arithmetic Mean

Ratings induce
various skewsymmetric matrices.

From David 1988 – The
Method of Paired Comparisons



David Gleich · Purdue

Netflix

28/40

Log-odds
Extracting the scores
Extracting the scores

do we have?

Do we trust all   
Not really.

David F. Gleich (Purdue)

105

101
101
105
Number of Comparisons

?
Netflix data 17k movies,
500k users, 100M ratings–
99.17% filled

KDD 2011
David

Gleich · Purdue

Netflix

29/40

How many   
Most.

107
Movie Pairs

Given    with all entries, then
  
is the Borda
count, the least-squares
solution to   

8/20
Only partial info? COMPLETE
IT!
Only partial info? Complete it!
Let   

be known for   

We trust these scores.

Goal Find the simplest skew-symmetric matrix that matches
the data   

  

noiseless

  
Both of these are NP-hard too.

David F. Gleich (Purdue)

David Gleich · Purdue

KDD 2011

Netflix

30/40

noisy

9/20
From a French nuclear test in 1970, imageNetflix
from http://picdit.wordpress.com/2008/07/21/8David Gleich · Purdue
insane-nuclear-explosions/

31/40

Solution GO NUCLEAR!
The ranking algorithm

The Ranking Algorithm
0. INPUT    (ratings data) and c
(for trust on comparisons)
1. Compute    from   
2. Discard entries with fewer than
c comparisons
3. Set   
to be indices and
values of what’s left
4.   

= SVP(  

)

David Gleich · Purdue

Netflix

32/40

5. OUTPUT   
Exact recovery
Exactrecovery
 results

Fraction of trials recovered

indices. Instead we view the following theorem as providing
intuition for the noisy problem.
Consider the operator basis for Hermitian matrices:

H = S [ K [ D where
p
S = {1/ 2(ei eT + ej eT ) : 1  i < j  n};
j
i
David Gross showed how to recover Hermitian matrices.
p
K = {ı/ 2(ei eT ej eT ) : 1we get n}; exact   
j
i
i.e. the conditions under which  i < j the

Note that   

D = {ei eT : 1  i  n}.
i

1
0.8
0.6
0.4
0.2
0
2
10

is Hermitian. Thus our new result!
T

Theorem 5. Let s be centered, i.e., s e = 0. Let Y =
seT
esT where ✓ = maxi s2 /(sT s) and ⇢ = ((maxi si )
i
(mini si ))/ksk. Also, let ⌦ ⇢ H be a random set of elements
with size |⌦| O(2n⌫(1 + )(log n)2 ) where ⌫ = max((n✓ +
1)/4, n⇢2 ). Then the solution of
minimize

kXk⇤

Figure
ity of
about
both th
§6.1 fo

6.1 R

The fi
subject to trace(X W i ) = trace((ıY ) W i ), W i 2 ⌦
ability o
the nois
is equal to ıY with probability at least 1 n .
with un
These a
The proof of this theorem follows directly by Theorem 4 if Netflix
 = se
David Gleich · Purdue
Y
  
⇤

33/40

⇤
Recovery Discussion and Experiments
Confession If   

, then just look at differences from
a connected set. Constants? Not very good.

  

Intuition for the truth.
  

David Gleich · Purdue

Netflix

34

  
Recovery Discussion and Experiments
Recovery Experiments
 look at differences from
Confession If   
, then just
a connected set. Constants? Not very good.

  

Intuition for the truth.
  

KDD 2011

16/20

David Gleich · Purdue

Netflix

35/40

David F. Gleich (Purdue)

  
Evaluation
Nuclear norm ranking

Mean rating
1

Median Kendall’s Tau

0.9
0.8
20
10
5
2
1.5

0.7
0.6
0.5

0.9
0.8
0.7
0.6
0.5

0

0.2

0.4 0.6
Error

0.8

1

0

0.2

0.4 0.6
Error

0.8

1

Figure 3: The performanceDavid Gleich · Purdue
of our algorithm Netflix
(left)

36/40

Median Kendall’s Tau

1
Tie in with PageRank
Another way to compute the scores is through a
close relative of PageRank and the linkprediction methods.


Massey or Colley methods
(2I + D A)s = “differeneces”
(L + 2D 1 )x = “scaled differences”

David Gleich · Purdue

Netflix

37/40
Ongoing Work
Finding communities in large networks !
We have the best community finder (as of CIKM2013)"
Whang, Gleich, Dhillon (CIKM)

Fast clique detection!
We have the fastest solver for max-clique problems, useful for
computing temporal strong components (Rossi, Gleich, et al. arXiv)

Scalable network alignment !

w
v
s

Overlap

r

& Low-rank clustering with features + links!
wtu

u

t

A

L

B

& Evolving network analysis!
David Gleich · Purdue

Netflix

38

& Scalable, distributed implementations !
of fast graph kernels!
References
!
Papers 
Gleich & Lim, KDD 2011 – Nuclear Norm Ranking"
Esfandiar, Gleich, Bonchi et al. – WAW2010, J. Internet. Math. 2013"
Kloster & Gleich, WAW2013, arXiv 1310.3423

Code!

www.cs.purdue.edu/homes/dgleich/codes!
bit.ly/dgleich-code


Supported by NSF CAREER 1149756-CCF 

www.cs.purdue.edu/homes/dgleich
David Gleich · Purdue
Netflix

39

!

!

More Related Content

PDF
Engineering Data Science Objectives for Social Network Analysis
PDF
Big data matrix factorizations and Overlapping community detection in graphs
PDF
Fast relaxation methods for the matrix exponential
PDF
Uncertainty in Deep Learning
PDF
The Lazy Traveling Salesman Memory Management for Large-Scale Link Discovery
PDF
Multimodal Residual Learning for Visual Question-Answering
PDF
EuroSciPy 2019 - GANs: Theory and Applications
PDF
Uncertainty Modeling in Deep Learning
Engineering Data Science Objectives for Social Network Analysis
Big data matrix factorizations and Overlapping community detection in graphs
Fast relaxation methods for the matrix exponential
Uncertainty in Deep Learning
The Lazy Traveling Salesman Memory Management for Large-Scale Link Discovery
Multimodal Residual Learning for Visual Question-Answering
EuroSciPy 2019 - GANs: Theory and Applications
Uncertainty Modeling in Deep Learning

What's hot (12)

PDF
IMAGE GENERATION WITH GANS-BASED TECHNIQUES: A SURVEY
PDF
Finding connections among images using CycleGAN
PDF
GAN in medical imaging
PPTX
Usage of Generative Adversarial Networks (GANs) in Healthcare
PPTX
ICASSP 2018 Tutorial: Generative Adversarial Network and its Applications to ...
PDF
GAN - Theory and Applications
PDF
Photo-realistic Single Image Super-resolution using a Generative Adversarial ...
PDF
Uncertainty Quantification in AI
PDF
[GAN by Hung-yi Lee]Part 1: General introduction of GAN
PPTX
Generative Adversarial Networks and Their Applications in Medical Imaging
PDF
Deep Convolutional GANs - meaning of latent space
PPTX
A (Very) Gentle Introduction to Generative Adversarial Networks (a.k.a GANs)
IMAGE GENERATION WITH GANS-BASED TECHNIQUES: A SURVEY
Finding connections among images using CycleGAN
GAN in medical imaging
Usage of Generative Adversarial Networks (GANs) in Healthcare
ICASSP 2018 Tutorial: Generative Adversarial Network and its Applications to ...
GAN - Theory and Applications
Photo-realistic Single Image Super-resolution using a Generative Adversarial ...
Uncertainty Quantification in AI
[GAN by Hung-yi Lee]Part 1: General introduction of GAN
Generative Adversarial Networks and Their Applications in Medical Imaging
Deep Convolutional GANs - meaning of latent space
A (Very) Gentle Introduction to Generative Adversarial Networks (a.k.a GANs)
Ad

Viewers also liked (20)

PDF
PageRank Centrality of dynamic graph structures
PDF
Overlapping clusters for distributed computation
PDF
Localized methods in graph mining
PDF
Graph libraries in Matlab: MatlabBGL and gaimc
PDF
Non-exhaustive, Overlapping K-means
PDF
Using Local Spectral Methods to Robustify Graph-Based Learning
PDF
Spacey random walks and higher order Markov chains
PDF
Iterative methods with special structures
PDF
Gaps between the theory and practice of large-scale matrix-based network comp...
PDF
A history of PageRank from the numerical computing perspective
PDF
A multithreaded method for network alignment
PDF
Direct tall-and-skinny QR factorizations in MapReduce architectures
PDF
What you can do with a tall-and-skinny QR factorization in Hadoop: Principal ...
PDF
Tall and Skinny QRs in MapReduce
PDF
Anti-differentiating approximation algorithms: A case study with min-cuts, sp...
PDF
MapReduce Tall-and-skinny QR and applications
PDF
Iterative methods for network alignment
PDF
Anti-differentiating Approximation Algorithms: PageRank and MinCut
PDF
The power and Arnoldi methods in an algebra of circulants
PDF
Relaxation methods for the matrix exponential on large networks
PageRank Centrality of dynamic graph structures
Overlapping clusters for distributed computation
Localized methods in graph mining
Graph libraries in Matlab: MatlabBGL and gaimc
Non-exhaustive, Overlapping K-means
Using Local Spectral Methods to Robustify Graph-Based Learning
Spacey random walks and higher order Markov chains
Iterative methods with special structures
Gaps between the theory and practice of large-scale matrix-based network comp...
A history of PageRank from the numerical computing perspective
A multithreaded method for network alignment
Direct tall-and-skinny QR factorizations in MapReduce architectures
What you can do with a tall-and-skinny QR factorization in Hadoop: Principal ...
Tall and Skinny QRs in MapReduce
Anti-differentiating approximation algorithms: A case study with min-cuts, sp...
MapReduce Tall-and-skinny QR and applications
Iterative methods for network alignment
Anti-differentiating Approximation Algorithms: PageRank and MinCut
The power and Arnoldi methods in an algebra of circulants
Relaxation methods for the matrix exponential on large networks
Ad

Similar to Fast matrix primitives for ranking, link-prediction and more (20)

PDF
Xavier amatriain, dir algorithms netflix m lconf 2013
PDF
MLConf - Emmys, Oscars & Machine Learning Algorithms at Netflix
PDF
Rank aggregation via nuclear norm minimization
DOCX
Mining Large Streams of User Data for PersonalizedRecommenda.docx
PDF
Practical Data Science Workshop - Recommendation Systems - Collaborative Filt...
PDF
Machine Learning at Netflix Scale
PPTX
Олександр Обєдніков “Рекомендательные системы”
PDF
Recommender Systems, Matrices and Graphs
PPTX
Lessons learnt at building recommendation services at industry scale
PPTX
lecture26-mf.pptx
PPTX
lecture244-mf.pptx
PPT
Download
PPT
Download
PPTX
Teacher training material
PPTX
Recommendation system
PDF
Modeling Social Data, Lecture 8: Recommendation Systems
PDF
Large Scale Recommendation: a view from the Trenches
PPT
Collaborative filtering using orthogonal nonnegative matrix
PDF
Recommender Systems with Implicit Feedback Challenges, Techniques, and Applic...
PDF
Qcon SF 2013 - Machine Learning & Recommender Systems @ Netflix Scale
Xavier amatriain, dir algorithms netflix m lconf 2013
MLConf - Emmys, Oscars & Machine Learning Algorithms at Netflix
Rank aggregation via nuclear norm minimization
Mining Large Streams of User Data for PersonalizedRecommenda.docx
Practical Data Science Workshop - Recommendation Systems - Collaborative Filt...
Machine Learning at Netflix Scale
Олександр Обєдніков “Рекомендательные системы”
Recommender Systems, Matrices and Graphs
Lessons learnt at building recommendation services at industry scale
lecture26-mf.pptx
lecture244-mf.pptx
Download
Download
Teacher training material
Recommendation system
Modeling Social Data, Lecture 8: Recommendation Systems
Large Scale Recommendation: a view from the Trenches
Collaborative filtering using orthogonal nonnegative matrix
Recommender Systems with Implicit Feedback Challenges, Techniques, and Applic...
Qcon SF 2013 - Machine Learning & Recommender Systems @ Netflix Scale

More from David Gleich (12)

PDF
Correlation clustering and community detection in graphs and networks
PDF
Spectral clustering with motifs and higher-order structures
PDF
Higher-order organization of complex networks
PDF
Spacey random walks and higher-order data analysis
PDF
Localized methods for diffusions in large graphs
PDF
Recommendation and graph algorithms in Hadoop and SQL
PDF
Personalized PageRank based community detection
PDF
Vertex neighborhoods, low conductance cuts, and good seeds for local communit...
PDF
A dynamical system for PageRank with time-dependent teleportation
PDF
How does Google Google: A journey into the wondrous mathematics behind your f...
PDF
Sparse matrix computations in MapReduce
PDF
Matrix methods for Hadoop
Correlation clustering and community detection in graphs and networks
Spectral clustering with motifs and higher-order structures
Higher-order organization of complex networks
Spacey random walks and higher-order data analysis
Localized methods for diffusions in large graphs
Recommendation and graph algorithms in Hadoop and SQL
Personalized PageRank based community detection
Vertex neighborhoods, low conductance cuts, and good seeds for local communit...
A dynamical system for PageRank with time-dependent teleportation
How does Google Google: A journey into the wondrous mathematics behind your f...
Sparse matrix computations in MapReduce
Matrix methods for Hadoop

Recently uploaded (20)

PDF
Peak of Data & AI Encore- AI for Metadata and Smarter Workflows
PPT
“AI and Expert System Decision Support & Business Intelligence Systems”
PPTX
VMware vSphere Foundation How to Sell Presentation-Ver1.4-2-14-2024.pptx
PDF
NewMind AI Monthly Chronicles - July 2025
PDF
Bridging biosciences and deep learning for revolutionary discoveries: a compr...
PDF
Reach Out and Touch Someone: Haptics and Empathic Computing
PDF
Agricultural_Statistics_at_a_Glance_2022_0.pdf
PDF
Review of recent advances in non-invasive hemoglobin estimation
PDF
KodekX | Application Modernization Development
PDF
Diabetes mellitus diagnosis method based random forest with bat algorithm
PDF
Building Integrated photovoltaic BIPV_UPV.pdf
PPTX
20250228 LYD VKU AI Blended-Learning.pptx
PDF
Unlocking AI with Model Context Protocol (MCP)
PDF
How UI/UX Design Impacts User Retention in Mobile Apps.pdf
PDF
TokAI - TikTok AI Agent : The First AI Application That Analyzes 10,000+ Vira...
PDF
Encapsulation_ Review paper, used for researhc scholars
PDF
cuic standard and advanced reporting.pdf
PDF
Architecting across the Boundaries of two Complex Domains - Healthcare & Tech...
PDF
Empathic Computing: Creating Shared Understanding
PPTX
Understanding_Digital_Forensics_Presentation.pptx
Peak of Data & AI Encore- AI for Metadata and Smarter Workflows
“AI and Expert System Decision Support & Business Intelligence Systems”
VMware vSphere Foundation How to Sell Presentation-Ver1.4-2-14-2024.pptx
NewMind AI Monthly Chronicles - July 2025
Bridging biosciences and deep learning for revolutionary discoveries: a compr...
Reach Out and Touch Someone: Haptics and Empathic Computing
Agricultural_Statistics_at_a_Glance_2022_0.pdf
Review of recent advances in non-invasive hemoglobin estimation
KodekX | Application Modernization Development
Diabetes mellitus diagnosis method based random forest with bat algorithm
Building Integrated photovoltaic BIPV_UPV.pdf
20250228 LYD VKU AI Blended-Learning.pptx
Unlocking AI with Model Context Protocol (MCP)
How UI/UX Design Impacts User Retention in Mobile Apps.pdf
TokAI - TikTok AI Agent : The First AI Application That Analyzes 10,000+ Vira...
Encapsulation_ Review paper, used for researhc scholars
cuic standard and advanced reporting.pdf
Architecting across the Boundaries of two Complex Domains - Healthcare & Tech...
Empathic Computing: Creating Shared Understanding
Understanding_Digital_Forensics_Presentation.pptx

Fast matrix primitives for ranking, link-prediction and more

  • 1. Fast matrix primitives for ranking, communities ! and more. David F. Gleich! David Gleich · Purdue Netflix 1 Computer Science! Purdue University!
  • 3. error 1 Models Previous work – and algorithms for high performance ! from the PI tackled net- computations matrix and network FIGURE 6 std 2 work alignment with matrix methods =for cm edge Std, s 0.39 Big data(b)methods overlap: SIMAX ‘09, SISC ‘11,MapReduce ‘11, ICASSP ’12 1 j i i0 Overlap Overlap j0 error SC ‘05, WAW ‘07, SISC ‘10, WWW ’10, … Massive matrix " computations std 0 0 Fast & Scalable" Network centrality 0 20 10 10 A L B Tensor eigenvalues" 0 (d) Std, s = 1.95 cm Ax = b min kAx bk Ax = x This proposal is for matchand a power method Network alignment tensor ing triangles using P methods: on multi-threaded maximize Tijk xi xj xk model compared to the prediction standard debble locations at the final time for two values of ICDM ‘09, SC ‘11, TKDE ‘13 = 1.95 cm. (Colors are visible in the electronic ijk n and kxk2 = 1 subject to distributed j Triangle j X i s i approximately twenty minutes to construct using (next) architectures k [x ]i = ⇢ · ( Tijk xj xk + xi ) k s. jk s- involved a few pre- and post-processing steps: ta where ! ensures the 2-norm m Aria, globally transpose the data, compute the g errors. The preprocessing steps took approx- SSHOPM method due to " nd 0 0 Data clustering WSDM ‘12, KDD ‘12, CIKM ’13 … 0 A recise timing information, L we do notB but report David Kolda and Mayo Gleich · Purdue Netflix 3 t r o s. g n. o
  • 4. The talk ends, you believe -- whatever you want to. Image from rockysprings, deviantart, CC share-alike 4 Everything in the world can be explained by a matrix, and we see how deep the rabbit hole goes
  • 5. Matrix computations in a red-pill David Gleich · Purdue Netflix 5 Solve a problem better by exploiting its structure!
  • 6. Problem 1 – (Faster) ! Recommendation as link prediction WHY NO PREPROCESSING? Top-k predicted “links” are movies to watch! David F. Gleich (Purdue) David Gleich · Purdue Emory Math/CS Seminar Netflix 6 Pairwise scores give user similarity 19 of 47
  • 7. David Gleich · Purdue Netflix 7 Problem 2 – (Better) ! Best movies
  • 8. Matrix computations in a red-pill David Gleich · Purdue Netflix 8 Solve a problem better by exploiting its structure!
  • 9. Matrix structure Netflix graph Movies " “liked” (>3 stars?) Problem 1! Adjacency matrix Normalized Laplacian matrix Random walk matrix Netflix matrix 1 1 4 5 5 Problem 2" Pairwise comparison matrix David Gleich · Purdue Netflix 9 5
  • 10. Problem 1 – (Faster) ! Recommendation as link prediction WHY NO PREPROCESSING? Top-k predicted “links” are movies to watch! David F. Gleich (Purdue) David Gleich · Purdue Emory Math/CS Seminar Netflix 10 Pairwise scores give user similarity 19 of 47
  • 11. z score (edge-based) is   movie  =   X  ↵   ✓                pred. on ` `=1 1 X num. paths of length ` from user to movie ✓ ◆ user k= (↵ A ) ind. vec. `=1 {z } | ` Math/CS Seminar Emory ` ⌘ei ◆ David Gleich · Purdue Netflix 11 1 Movie prediction" vector ) Matrix based link predictors
  • 12. Matrix based link predictors 1 X ✓ ◆ user k= (↵ A ) ind. vec. `=1 {z } | ` ` ⌘ei Neumann Carl Neumann 1 X k =0 ↵A)k = ei (tA)k David Gleich · Purdue Netflix 12 (I
  • 13. The Katz score (edge-based) is Matrix based link predictors                               (I ↵A)k = ei PageRank (I ↵P)x = ei Semi-super." learning Heat kernel (IDavid F. Gleich (Purdue) = ei ↵L)x exp{↵P}x = ei Emory Math/CS Seminar They all look at sums of " damped paths, but " change the details, slightly David Gleich · Purdue Netflix 13 Katz
  • 14. Matrix based link predictors are localized! PageRank scores for one node! Crawl of flickr from 2006 ~800k nodes, 6M edges, alpha=1/2 0 1.5 error ||xtrue – xnnz||1 10 1 0.5 −10 10 −15 0 2 4 plot(x) 6 8 10 5 x 10 10 0 10 2 4 10 6 10 10 nonzeros David Gleich · Purdue Netflix 14 0 −5 10
  • 15. Matrix based link predictors are localized! KATZ SCORES ARE LOCALIZED David F. Gleich (Purdue) Emory Math/CS Seminar 32 of 47 David Gleich · Purdue Netflix 15 Up to 50 neighbors is 99.65% of the total mass
  • 16. Matrix computations in a red-pill David Gleich · Purdue Netflix 16 Solve a problem better by exploiting its structure!
  • 17. How do we compute them fast? PageRank xj = ↵ X i neigh. of j xi deg(i) + 1 if j is the target user w/ access to in-links & degs. w/ access to out-links PageRankPull PageRankPush xj(k+1) (k) ↵xa /6 (k) ↵xb /2 (k) ↵xc /3 = fj xj(k+1) ↵ X i!j xi(k ) /degi = fj Let b a c j = blue node (k+1) = xj(k) + rj (k +1) =0 then xj Update r(k +1) rj (k (k) ra +1) = ra + ↵rj(k ) /3 (k (k) rb +1) = rb + ↵rj(k ) /3 (k (k) rc +1) = rc + ↵rj(k ) /3 David Gleich · Purdue Netflix 17 (k +1) Solve for xj j = blue node
  • 18. We have good theory for this algorithm … David Gleich · Purdue Netflix 18 … and even better empirical performance.
  • 19. Theory Andersen, Chung, Lang (2006)! For PageRank, “fast runtimes” and “localization” Bonchi, Esfandiar, Gleich, et al. (2010/2013)! For Katz, “fast runtimes” David Gleich · Purdue Netflix 19 Kloster, Gleich (2013)! For Katz, Heat Kernel, " “fast runtimes” and “localization”" (assuming power-law degrees)
  • 20. Accuracy vs. work ! (Heat kernel) dblp−cc dblp collaboration graph, 225k vertices 1 0.6 tol=10−5 tol=10−4 0.4 @10 @25 0.2 @100 @1000 0 −2 −1 0 10 10 10 Effective matrix−vector products David Gleich · Purdue Netflix 20 Precision 0.8 For the dblp collaboration graph, we study the precision in finding the 100 largest nodes as we vary the work. This set of 100 does not include the nodes immediate neighbors. (One column, but representative)
  • 21. David Gleich · Purdue Netflix 21 Empirical runtime (Katz) TIMING
  • 22. 40 Never got to try it … analytics test on HelloMovies.com Need to ix now matr Netflix 60 80 Ran out of money once we had the algorithms … promising initial results though! I collaborate with the company behind He David Gleich · Purdue Netflix 22 Note 1 1 1 1 1 1 1 1
  • 23. David Gleich · Purdue Netflix 23 Problem 2 – (Better) ! Best movies
  • 24. Which is a better list of good DVDs? Lord of the Rings 3: The Return of … Lord of the Rings 3: The Return of … Lord of the Rings 1: The Fellowship Lord of the Rings 1: The Fellowship Lord of the Rings 2: The Two Towers Lord of the Rings 2: The Two Towers Lost: Season 1 Star Wars V: Empire Strikes Back Battlestar Galactica: Season 1 Raiders of the Lost Ark Fullmetal Alchemist Star Wars IV: A New Hope Trailer Park Boys: Season 4 Shawshank Redemption Trailer Park Boys: Season 3 Star Wars VI: Return of the Jedi Tenchi Muyo! Lord of the Rings 3: Bonus DVD Shawshank Redemption The Godfather Nuclear Norm " based rank aggregation (the mean rating) (not matrix completion on the netflix rating matrix) David Gleich · Purdue Netflix 24/40 Standard " rank aggregation"
  • 25. Rank Aggregation Given partial orders on subsets of items, rank aggregation is the problem of finding an overall ordering. Voting Find the winning candidate Program committees Find the best papers given reviews Dining Find the best restaurant in Chicago David Gleich · Purdue Netflix 25/40
  • 26. Ranking is really hard John Kemeny Ken Arrow All rank aggregations involve some measure of compromise A good ranking is the “average” ranking under a permutation distance NP hard to compute Kemeny’s ranking David Gleich · Purdue Netflix 26/40 Dwork, Kumar, Naor, ! Sivikumar
  • 27. Supposewe had scores Suppose we had scores Let    be the score of the ith movie/song/paper/team to rank Suppose we can compare the ith to jth:    is skew-symmetric, rank 2. Also works for    with an extra log. Numerical ranking is intimately intertwined with skew-symmetric matrices Kemeny and Snell, Mathematical Models in Social Sciences (1978) David F. Gleich (Purdue) KDD 2011 David Gleich · Purdue Netflix 6/20 27/40 Then   
  • 28. Using ratings as comparisons Arithmetic Mean Ratings induce various skewsymmetric matrices. From David 1988 – The Method of Paired Comparisons David Gleich · Purdue Netflix 28/40 Log-odds
  • 29. Extracting the scores Extracting the scores do we have? Do we trust all    Not really. David F. Gleich (Purdue) 105 101 101 105 Number of Comparisons ? Netflix data 17k movies, 500k users, 100M ratings– 99.17% filled KDD 2011 David Gleich · Purdue Netflix 29/40 How many    Most. 107 Movie Pairs Given    with all entries, then    is the Borda count, the least-squares solution to    8/20
  • 30. Only partial info? COMPLETE IT! Only partial info? Complete it! Let    be known for    We trust these scores. Goal Find the simplest skew-symmetric matrix that matches the data       noiseless    Both of these are NP-hard too. David F. Gleich (Purdue) David Gleich · Purdue KDD 2011 Netflix 30/40 noisy 9/20
  • 31. From a French nuclear test in 1970, imageNetflix from http://picdit.wordpress.com/2008/07/21/8David Gleich · Purdue insane-nuclear-explosions/ 31/40 Solution GO NUCLEAR!
  • 32. The ranking algorithm The Ranking Algorithm 0. INPUT    (ratings data) and c (for trust on comparisons) 1. Compute    from    2. Discard entries with fewer than c comparisons 3. Set    to be indices and values of what’s left 4.    = SVP(   ) David Gleich · Purdue Netflix 32/40 5. OUTPUT   
  • 33. Exact recovery Exactrecovery results Fraction of trials recovered indices. Instead we view the following theorem as providing intuition for the noisy problem. Consider the operator basis for Hermitian matrices: H = S [ K [ D where p S = {1/ 2(ei eT + ej eT ) : 1  i < j  n}; j i David Gross showed how to recover Hermitian matrices. p K = {ı/ 2(ei eT ej eT ) : 1we get n}; exact    j i i.e. the conditions under which  i < j the Note that    D = {ei eT : 1  i  n}. i 1 0.8 0.6 0.4 0.2 0 2 10 is Hermitian. Thus our new result! T Theorem 5. Let s be centered, i.e., s e = 0. Let Y = seT esT where ✓ = maxi s2 /(sT s) and ⇢ = ((maxi si ) i (mini si ))/ksk. Also, let ⌦ ⇢ H be a random set of elements with size |⌦| O(2n⌫(1 + )(log n)2 ) where ⌫ = max((n✓ + 1)/4, n⇢2 ). Then the solution of minimize kXk⇤ Figure ity of about both th §6.1 fo 6.1 R The fi subject to trace(X W i ) = trace((ıY ) W i ), W i 2 ⌦ ability o the nois is equal to ıY with probability at least 1 n . with un These a The proof of this theorem follows directly by Theorem 4 if Netflix = se David Gleich · Purdue Y    ⇤ 33/40 ⇤
  • 34. Recovery Discussion and Experiments Confession If    , then just look at differences from a connected set. Constants? Not very good.    Intuition for the truth.    David Gleich · Purdue Netflix 34   
  • 35. Recovery Discussion and Experiments Recovery Experiments look at differences from Confession If    , then just a connected set. Constants? Not very good.    Intuition for the truth.    KDD 2011 16/20 David Gleich · Purdue Netflix 35/40 David F. Gleich (Purdue)   
  • 36. Evaluation Nuclear norm ranking Mean rating 1 Median Kendall’s Tau 0.9 0.8 20 10 5 2 1.5 0.7 0.6 0.5 0.9 0.8 0.7 0.6 0.5 0 0.2 0.4 0.6 Error 0.8 1 0 0.2 0.4 0.6 Error 0.8 1 Figure 3: The performanceDavid Gleich · Purdue of our algorithm Netflix (left) 36/40 Median Kendall’s Tau 1
  • 37. Tie in with PageRank Another way to compute the scores is through a close relative of PageRank and the linkprediction methods. Massey or Colley methods (2I + D A)s = “differeneces” (L + 2D 1 )x = “scaled differences” David Gleich · Purdue Netflix 37/40
  • 38. Ongoing Work Finding communities in large networks ! We have the best community finder (as of CIKM2013)" Whang, Gleich, Dhillon (CIKM) Fast clique detection! We have the fastest solver for max-clique problems, useful for computing temporal strong components (Rossi, Gleich, et al. arXiv) Scalable network alignment ! w v s Overlap r & Low-rank clustering with features + links! wtu u t A L B & Evolving network analysis! David Gleich · Purdue Netflix 38 & Scalable, distributed implementations ! of fast graph kernels!
  • 39. References ! Papers Gleich & Lim, KDD 2011 – Nuclear Norm Ranking" Esfandiar, Gleich, Bonchi et al. – WAW2010, J. Internet. Math. 2013" Kloster & Gleich, WAW2013, arXiv 1310.3423 Code! www.cs.purdue.edu/homes/dgleich/codes! bit.ly/dgleich-code Supported by NSF CAREER 1149756-CCF www.cs.purdue.edu/homes/dgleich David Gleich · Purdue Netflix 39 ! !