SlideShare a Scribd company logo
Latent Interest and Topic Mining on User-item
Bipartite Networks
Jinliang Xu, Shangguang Wang∗,Sen Su
State Key Laboratory of Networking and Switching Technology
Beijing University of Posts and Telecommunications,
Beijing, China
{jlxu,sgwang,susen}@bupt.edu.cn
Sathish A.P Kumar
Department of Computer Science and Information Systems, Coastal Carolina University
South Carolina, USA
skumar@coastal.edu
Wu Chou
Huawei Technologies Co., Ltd
Shenzhen, China
Wu.Chou@huawei.com
Abstract—Latent Factor Model (LFM) is extensively used in
dealing with user-item bipartite networks in service recom-
mendation systems. To alleviate the limitations of LFM, this
papers presents a novel unsupervised learning model, Latent
Interest and Topic Mining model (LITM), to automatically
mine the latent user interests and item topics from user-item
bipartite networks. In particular, we introduce the motivation
and objectives of this bipartite network based approach, and
detail the model development and optimization process of the
proposed LITM. This work not only provides an efficient method
for latent user interest and item topic mining, but also highlights
a new way to improve the accuracy of service recommendation.
Experimental studies are performed and the results validate the
LITM’s efficiency in model training, and its ability to provide
better service recommendation performance based on user-item
bipartite networks are demonstrated.
Keywords—service recommendation, user-item bipartite net-
work, latent topic and interest, interpretability, and efficiency.
I. INTRODUCTION & RELATED WORK
Service recommendation systems collect information on
the preferences of its users for a set of items to help users
identify useful items from a considerably large search space
(e.g.,Taobao Mall, Netflix movies). With the rapid develop-
ment of e-commerce and mobile internet, new challenges have
been posed to researchers who are trying to mine the inner
patterns about users, items and relationship between them,
such as exploiting the purchasing behaviors in order to identify
additional possible matches, i.e. connecting users with items
they would otherwise not be able to find.
In the study of mining the relationships between users and
items, an important and special class of networks is a bipartite
network, where nodes can be divided into two disjoint sets,
such that no two nodes within the same set are connected [1],
[2], [3]. Many systems can be naturally modeled as bipartite
networks. For example, users are connected with the movies
they have rated (e.g., MovieLens [4]), web-users are connected
with the web sites that they have collected in their bookmark
sites, and the consumers are connected with the goods that
they have purchased from the market.
In dealing with the user-item bipartite networks, the tra-
ditional Latent Factor Models (LFM) have been shown to
be effective [5], [6]. LFM is able to factor the user-item
bipartite networks into two smaller matrices, consisting of
latent user factor vectors and item factor vectors respectively.
However, the output latent factor vectors generated by LFM
may contain arbitrary values, including negative numbers,
causing low interpretability [7]. As a result, LFM is hard to
explain to its users how a specific recommendation is derived.
Furthermore, each of the output latent factor vectors are not
a probability vector (Note that a probability vector here is
a vector with non-negative elements that add up to 1), this
property makes LFM difficult for building more complex,
realistic models [8]. Similar approaches to recommendation
systems include LSA [9], and NMF [10]. On the other hand,
LDA [11], [12] can output probabilistic distribution of latent
topics when taking user-item bipartite network as input, and
it has better interpretability than LFM. However, LDA based
models cannot model both users and items [12]. In addition,
LDA based models may not be applied directly to mine user
interests [2]. For example, Haveliwala et. al [13] computed a
set of scores for each item, where each score is related to a
topic. However, the topics here can not be automatically mined
from the user-item bipartite network.
In this paper, by combining the best of LFM and LDA,
we propose a novel unsupervised learning model, named
Latent Interest and Topic Mining model (LITM) to model both
users and items with the better interpretability from user-item
bipartite networks. As shown in Fig. 1, LITM is based on
the following: 1) for each item there is a probability vector
2016 IEEE International Conference on Services Computing
978-1-5090-2628-9/16 $31.00 © 2016 IEEE
DOI 10.1109/SCC.2016.105
778
Fig. 1. Concept of latent user interests and item topics in LITM. At the left,
it is a typical user-item bipartite network in recommender system, where a
user is connected with an item if he has purchased it. The latent user interest
and item topic vector are shown at the right side.
called latent item topic vector; 2) for each user there is a same
length vector called latent user interest vector, in which each
position of the latter vector corresponds to the same position of
the former vector; and 3) the occurrence probability of each
connection is determined by the corresponding user’s latent
interest vector and the corresponding item’s latent topic vector.
Latent user interest vector and latent item topic vector can
generalize the intrinsic patterns underlying the data generation
process into smaller number of latent factors. Taking the
MovieLens as an example, each movie has a probability dis-
tribution over different topics, and each user has a probability
distribution of interests over all topics.
The remainder of this paper is organized as follows. In
Section II we detail the model development and optimization
process of the proposed LITM. In Section III, our experimental
studies show that the proposed LITM is more efficient in mod-
el training against the baseline and it improves the accuracy of
service recommendation. Finally, we summarize the findings
of this paper in Section IV.
II. THE PROPOSED MODEL LITM
In this section we provide the details of the model devel-
opment and optimization process of LITM.
A. Conception and Notations
We explicitly describe the concept of latent item topic
vectors and latent user interest vectors respectively by the
following:
• latent item topic vector vm(m = 1, 2, · · · , M): for each
item m, we use K-dimensionality probability vector vm
to represent its topic probability distribution over K ∈ Z+
topics;
• latent user interest vector un(n = 1, 2, · · · N): for each
user n, we use the same length probability vector un
to represent the probability distribution of user’s interest
over the K topics. Each position of un corresponds to
the same position in vm.
Then we define the occurrence probability of the connection
between user n and item m (denoted by un − vm), or the
probability that user n select item m, by the following:
uT
n vm =
K
k=1
un,kvm,k, (1)
where un,k and vm,k are respectively the k−th elements of
un and vm. By doing so, we let un,kvm,k be the occurrence
probability of un −vm on the k−th topic, then the sum of the
occurrence probability over all topics (Eq. 1) can represent the
occurrence probability of the connection un − vm.
B. LITM
Suppose both un(n = 1, 2, · · · N) and vm(m =
1, 2, · · · , M) are known, the occurrence of all connections
un − vm in bipartite network G are independent, according to
the d-separation criterion in probabilistic graphical models.
So we can define the objective function as the occurrence
probability of the whole bipartite network G as
P(G|U, V ) =
N
n=1
M
m=1
un
T
vm
σm,n
, (2)
where σm,n indicates whether the connection un − vm exists
(σm,n = 1) or not (σm,n = 0), and U = {un|n =
1, 2, · · · , N}, V = {vm|m = 1, 2, · · · , M}.
For the ease of optimization, we take the negative logarithm
of the objective function, then add two regularized items, and
we get the problem to be optimized by the following:
minU,V −
N
n=1
M
m=1
σm,n log un
T
vm +
1
N
N
n=1
||un||2 +
1
M
M
m=1
||vm||2
s.t. un ≥ 0, ||un||1 = 1, n = 1, 2, · · · , N
vm ≥ 0, ||vm||1 = 1, m = 1, 2, · · · , M
. (3)
Note that we add the latter two 2 norm regularized terms to
alleviate the potential overfitting.
Analyzing the above problems shows, it shows as follows:
1) This objective function is non-convex (for the un
T
vm
term). Note, however, if we fix U and treat it as constants,
then the objective is a convex function of V , and vice versa.
We will therefore fix V and optimize U, then fix U and
optimize V , thereby obtaining a series of efficiently solvable
subproblems. As each update of U and V , it decreases the
objective function, from which the process will converge and
find a local minimum eventually. 2) If we fix V and optimize
U, then un, n = 1, 2, · · · , N can be optimized individually,
and vice versa. We take optimizing un as an example, then
we get a subproblem like the following:
minun −
M
m=1
σm,n log un
T
vm +
1
N
||un||2
s.t. un ≥ 0, ||un||1 = 1
, (4)
where vm ≥ 0, ||vm||1 = 1, m = 1, 2, · · · , M.
Clearly the subproblem (Eq. 4) has no closed-form solution,
and cannot use the gradient descent method directly to get the
optimal numerical solution for the given constraints un ≥
0, ||un||1 = 1. However, since this is a convex programming
779
problem with linear constraints and according to the convexity-
preserving rules, we can apply Reduced Gradient method
to search for an optimal solution for the abovementioned
subproblems.
Algorithm 1: The adaptive Reduced Gradient method
Input: σm,n ∈ {0, 1}, vm, m = 1, 2, · · · , M; K ∈ Z+
;
Output: u
1 Initialize u > 0 ∧ ||u||1 = 1, d = 0;
2 while d = 0 do
3 umax = max{u};
4 uN = uumax //uN contains all but umax;
5 g = −
M
m=1 σm,n
vm
uT vm
+ 2
N un//gradient vector ;
6 gmax = gj|uj =umax
// the gradient value at umax;
7 for j = 1; j ≤ K; j + + do
8 if uj ∈ uN then
9 rj = gj − gmax;
10 else
11 loc = j;
12 rj = 0;
13 for j = 1; j ≤ K; j + + do
14 if rj > 0 then
15 dj = −ujrj;
16 else
17 dj = −rj;
18 dloc = −sum{d} //sum{d} is sum value of d;
19 λmax = ∞;
20 for j = 1; j ≤ K; j + + do
21 if dj < 0 then
22 λmax = min λmax, −
uj
dj
23 search the best step length λbest between [0, λmax];
24 u = u + λbestd;
25 return u;
The Reduced Gradient method uses the equality constraints
to eliminate a subset of variables, thereby it reduces the orig-
inal problem to a bound-constrained problem in the space of
the remaining variables. Algorithm. 1 lists the whole process
to get an optimal solution for the subproblem (Eq. 4) using
the adaptive Reduced Gradient method. Lines 3-4 select the
largest variables as the basic variables. Lines 9-14 ensure
the feasibility of the new solution. Lines 15-19 make sure
to move in the direction to decrease the objective function.
Lines 21-24 calculate the largest possible step length towards
the optimizing direction.
Based on Algorithm. 1 and the previous analysis, we
summarize the basic idea of optimizing the proposed LITM
(Eq. 3) in Algorithm. 2, where lines 3-4 means fixing V and
optimizing U, line 5-6 means fixing U and optimizing V , and
line 2 means repeating the two operations until convergence.
This optimization process will lead us to a local minimum of
the objective function.
Algorithm 2: Basic idea for optimizing LITM
Input: σm,n ∈ {0, 1}, n = 1, 2, · · · , N, m =
1, 2, · · · , M; K ∈ Z+
;
Output: un, vm, n = 1, 2, · · · , N, m = 1, 2, · · · , M;
1 Initialize un and vm;
2 while not converged do
3 for n = 1; n ≤ N; n + + do
4 un =
minun −
M
m=1 σm,n log un
T
vm + 1
N ||un||2
// get the solution by Reduced Gradient
method(see Alg.1);
5 for m = 1; m ≤ M; m + + do
6 vm =
minvm −
N
n=1 σm,n log un
T
vm + 1
M ||vm||2
// get the solution by Reduced Gradient
method(see Alg.1);
7 return un, vm, n = 1, 2, · · · , N, m = 1, 2, · · · , M;
III. EXPERIMENTAL EVALUATION
In this section, we describe a series of experiments that we
conducted on benchmarking MovieLens dataset to evaluate
our proposed LITM and explore how much it can improve
the effectiveness of the service recommendation as well as its
efficiency in model training.
A. Experimental DataSet & Baseline Model
The MovieLens data is from the GroupLens Research1
. It
comprises the ratings on 1, 682 items (movies) and 943 users,
where each user votes on movies in five discrete ratings 1 ∼ 5.
To generate the user-item bipartite network, we consider that
the connection between an item and a user exists if and only
if the given rating is at least 3. By this way, we obtained a
82, 520 entry (an entry means an actual user-item connection)
user-item bipartite network as our experimental dataset.
In our experiments, the dataset was randomly partitioned
into two parts, where the training set contained 90% of the
total entries, and the remaining 10% (8,252 entries) entries
constitute the test set.
In regards to the baseline model (we set K = 19 currently,
and more work is under way), we preliminarily adopted the
traditional model LFM [5] that has been described in Section
I. We choosed LFM because it has been extensively used in
the domain of service recommendation, and similar to our
proposed LITM, in LFM each users and items has a latent
vector representation.
B. Experimental Results & Discussions
LFM based recommender systems can provide each user
with an ordered queue of all its uncollected items. Similarly
our proposed LITM can rank a user’s uncollected items in the
order of the occurrence probability of the connections using
1http://www.grouplens.org
780
Fig. 2. The predicted position of each entry in the
test data ranked in ascending order.
Fig. 3. The hitting rate as a function of the length
of recommendation list.
Fig. 4. Learning curves for LFM and LITM model
training.
Eq. 1. For a user un, if the connection un − vm is in the
test set, we measure the position of vm in the ordered queue.
For example, if there are 1000 uncollected items for un, and
vm is the 40th from the top, we say the position of vm is
the top 40/1000, denoted by rankn,m = 0.04. Since the test
entries are actual connections, a good model is expected to
give high recommendations to them, indicated by their good
rank order positions. Fig. 2 shows the position values of 8, 252
test entries, which are ordered from the top position (rank →
0) to the bottom position (rank → 1). Clearly, the proposed
LITM performs better than LFM.
To evaluate the proposed LITM in recommender systems
from another point of view, we adopted a measure of recom-
mending accuracy that depends on the length of recommenda-
tion list. The recommendation list for a user un with length L,
contains L highest ranking items generated from the model.
For each entry un − vm in the test data, if vm is in ui’s
recommendation list, we say the entry un −vm is “hit” by the
model. The ratio of hit entries is called “hitting rate”. For a
given L, the model with a higher hitting rate is better, and vice
versa. Clearly, the hitting rate monotonously increases with L,
with the upper bound 1 for sufficiently large L. In Fig. 3, we
report the hitting rate as a function of L for LFM and LITM,
and it’s easy to see that LITM had a better performance than
LFM.
We also evaluated the efficiency of model training. Both
training processes of LFM and LITM will generate a series
of solutions in model training, and we plotted the base-10
logarithm of the objective function values of these solutions in
Fig. 4. It shows that the proposed LITM converged faster than
the traditional LFM, which demonstrates LITM’s efficiency
in model training. By changing the learning rate in model
training, we obtained the similar results as described above.
Despite three distinct metrics being used in the performance
evaluation experiments using three distinct metrics, it consis-
tently demonstrated the LITM’s efficiency in model training
and its ability for better service recommendation performance.
IV. CONCLUSION
This paper presented a novel unsupervised learning model
LITM to automatically mine the latent user interest and item
topic distribution from user-item bipartite networks. LITM
improves LFM in its low interpretability. In addition, this
work not only provides an efficient method for latent user
interest and item topic mining, but also highlights a new way
to improve accuracy of service recommendation.
ACKNOWLEDGMENTS
The work presented in this study is supported by NSFC
(61472047); NSFC(61571066).
REFERENCES
[1] T. Hu, H. Xiong, and S. Y. Sung, “Co-preserving patterns in bipartite
partitioning for topic identification.,” in SIAM International Conference
on Data Mining (SDM), pp. 509–514, SIAM, 2007.
[2] X. Tang, M. Zhang, and C. C. Yang, “User interest and topic detec-
tion for personalized recommendation,” in Proceedings of the 2012
IEEE/WIC/ACM International Joint Conferences on Web Intelligence
and Intelligent Agent Technology (WI-IAT), pp. 442–446, IEEE Com-
puter Society, 2012.
[3] T. de Paulo Faleiros and A. de Andrade Lopes, “Bipartite graph for topic
extraction,” in Proceedings of the 24th International Joint Conference
on Artificial Intelligence (IJCAI), pp. 4361–4362, AAAI Press, 2015.
[4] F. M. Harper and J. A. Konstan, “The movielens datasets: History and
context,” ACM Transactions on Interactive Intelligent Systems, vol. 5,
no. 4, p. 19, 2015.
[5] R. Bell, Y. Koren, and C. Volinsky, “Modeling relationships at multiple
scales to improve accuracy of large recommender systems,” in Proceed-
ings of the 13th ACM SIGKDD International Conference on Knowledge
Discovery and Data Mining, pp. 95–104, ACM, 2007.
[6] Y. Shen and R. Jin, “Learning personal+ social latent factor model
for social recommendation,” in Proceedings of the 18th ACM SIGKDD
International Conference on Knowledge Discovery and Data Mining,
pp. 1303–1311, ACM, 2012.
[7] G. Friedrich and M. Zanker, “A taxonomy for generating explanations
in recommender systems,” AI Magazine, vol. 32, no. 3, pp. 90–98, 2011.
[8] S. R. Eddy, “A probabilistic model of local sequence alignment that sim-
plifies statistical significance estimation,” PLOS Computational Biology,
vol. 4, no. 5, p. e1000069, 2008.
[9] T. Hofmann, “Probabilistic latent semantic indexing,” in Proceedings of
the 22nd Annual International ACM SIGIR Conference on Research and
Development in Information Retrieval, pp. 50–57, ACM, 1999.
[10] D. D. Lee and H. S. Seung, “Algorithms for non-negative matrix
factorization,” in Proceedings of the Annual Conference on Neural
Information Processing Systems (NIPS), pp. 556–562, 2001.
[11] D. M. Blei, A. Y. Ng, and M. I. Jordan, “Latent dirichlet allocation,”
Journal of Machine Learning Research, vol. 3, pp. 993–1022, 2003.
[12] J. Tang, R. Jin, and J. Zhang, “A topic modeling approach and its
integration into the random walk framework for academic search,” in
Proceedings of the 8th IEEE International Conference on Data Mining
(ICDM), pp. 1055–1060, IEEE, 2008.
[13] T. H. Haveliwala, “Topic-sensitive pagerank: A context-sensitive ranking
algorithm for web search,” IEEE Transactions on Knowledge and Data
Engineering, vol. 15, no. 4, pp. 784–796, 2003.
781

More Related Content

PDF
Paper id 252014146
PDF
Finite_Element_Analysis_with_MATLAB_GUI
PDF
Centrality Prediction in Mobile Social Networks
PPTX
07 Statistical approaches to randomization
DOCX
StrucA final report
PDF
Review and comparison of tasks scheduling in cloud computing
PPT
Study on Some Key Issues of Synergetic Neural Network
PDF
Formulation of modularity factor for community detection applying
Paper id 252014146
Finite_Element_Analysis_with_MATLAB_GUI
Centrality Prediction in Mobile Social Networks
07 Statistical approaches to randomization
StrucA final report
Review and comparison of tasks scheduling in cloud computing
Study on Some Key Issues of Synergetic Neural Network
Formulation of modularity factor for community detection applying

What's hot (15)

DOCX
Learning to rank image tags with limited training examples
PDF
INTRA BLOCK AND INTER BLOCK NEIGHBORING JOINT DENSITY BASED APPROACH FOR JPEG...
PDF
S-CUBE LP: Service Discovery and Task Models
PPTX
Recommendation system
PDF
NEW ONTOLOGY RETRIEVAL IMAGE METHOD IN 5K COREL IMAGES
PDF
A Blind Steganalysis on JPEG Gray Level Image Based on Statistical Features a...
DOCX
LEARNING TO RANK IMAGE TAGS WITH LIMITED TRAINING EXAMPLES - IEEE PROJECTS I...
PDF
08 Exponential Random Graph Models (ERGM)
DOCX
Learning to rank image tags with limited training examples
PDF
IRJET- Image Captioning using Multimodal Embedding
PDF
The Validity of CNN to Time-Series Forecasting Problem
PPTX
Session-Based Recommendations with Recurrent Neural Networks (Balazs Hidasi, ...
PDF
Machine Learning Algorithms for Image Classification of Hand Digits and Face ...
PDF
An experimental evaluation of similarity-based and embedding-based link predi...
DOCX
Hidden Markov Model Classification And Gabor Filter Preprocessing
Learning to rank image tags with limited training examples
INTRA BLOCK AND INTER BLOCK NEIGHBORING JOINT DENSITY BASED APPROACH FOR JPEG...
S-CUBE LP: Service Discovery and Task Models
Recommendation system
NEW ONTOLOGY RETRIEVAL IMAGE METHOD IN 5K COREL IMAGES
A Blind Steganalysis on JPEG Gray Level Image Based on Statistical Features a...
LEARNING TO RANK IMAGE TAGS WITH LIMITED TRAINING EXAMPLES - IEEE PROJECTS I...
08 Exponential Random Graph Models (ERGM)
Learning to rank image tags with limited training examples
IRJET- Image Captioning using Multimodal Embedding
The Validity of CNN to Time-Series Forecasting Problem
Session-Based Recommendations with Recurrent Neural Networks (Balazs Hidasi, ...
Machine Learning Algorithms for Image Classification of Hand Digits and Face ...
An experimental evaluation of similarity-based and embedding-based link predi...
Hidden Markov Model Classification And Gabor Filter Preprocessing
Ad

Similar to Latent Interest and Topic Mining on User-item Bipartite Networks (20)

PDF
Nbe rtopicsandrecomvlecture1
PPTX
Recommenders, Topics, and Text
PDF
Recommender Systems In Industry
PDF
Latent Structured Ranking
PDF
Recent advances in deep recommender systems
PDF
Discovering User's Topics of Interest in Recommender Systems @ Meetup Machine...
PDF
Modeling Social Data, Lecture 8: Recommendation Systems
PDF
Discovering User's Topics of Interest in Recommender Systems
PDF
Inferring Complementary Products from Baskets and Browsing Sessions
PDF
Matrix Factorization Techniques For Recommender Systems
PDF
IRJET- Online Course Recommendation System
PDF
LatentCross.pdf
PDF
IRJET- Boosting Response Aware Model-Based Collaborative Filtering
PDF
Factorization Models For Multirelational Data 1st Edition Lucas Drumond
PDF
2017 10-10 (netflix ml platform meetup) learning item and user representation...
PPT
Recommender Systems Tutorial (Part 2) -- Offline Components
PPTX
Exploration of the Hidden Influential Factors on Crime.pptx
PPTX
Recommender Systems: Advances in Collaborative Filtering
PDF
Ire major project
PDF
Past, present, and future of Recommender Systems: an industry perspective
Nbe rtopicsandrecomvlecture1
Recommenders, Topics, and Text
Recommender Systems In Industry
Latent Structured Ranking
Recent advances in deep recommender systems
Discovering User's Topics of Interest in Recommender Systems @ Meetup Machine...
Modeling Social Data, Lecture 8: Recommendation Systems
Discovering User's Topics of Interest in Recommender Systems
Inferring Complementary Products from Baskets and Browsing Sessions
Matrix Factorization Techniques For Recommender Systems
IRJET- Online Course Recommendation System
LatentCross.pdf
IRJET- Boosting Response Aware Model-Based Collaborative Filtering
Factorization Models For Multirelational Data 1st Edition Lucas Drumond
2017 10-10 (netflix ml platform meetup) learning item and user representation...
Recommender Systems Tutorial (Part 2) -- Offline Components
Exploration of the Hidden Influential Factors on Crime.pptx
Recommender Systems: Advances in Collaborative Filtering
Ire major project
Past, present, and future of Recommender Systems: an industry perspective
Ad

More from jins0618 (20)

PDF
Machine Status Prediction for Dynamic and Heterogenous Cloud Environment
PDF
Web Service QoS Prediction Approach in Mobile Internet Environments
PDF
吕潇 星环科技大数据技术探索与应用实践
PPT
李战怀 大数据环境下数据存储与管理的研究
PPTX
2015 07-tuto0-courseoutline
PDF
Christian jensen advanced routing in spatial networks using big data
PPTX
Jeffrey xu yu large graph processing
PDF
Calton pu experimental methods on performance in cloud and accuracy in big da...
PDF
Ling liu part 02:big graph processing
PDF
Ling liu part 01:big graph processing
PPTX
Wang ke mining revenue-maximizing bundling configuration
PDF
Wang ke classification by cut clearance under threshold
PPTX
2015 07-tuto2-clus type
PPTX
2015 07-tuto1-phrase mining
PPTX
2015 07-tuto3-mining hin
PPTX
2015 07-tuto0-courseoutline
PPTX
Weiyi meng web data truthfulness analysis
PPTX
Ke yi small summaries for big data
PDF
Gao cong geospatial social media data management and context-aware recommenda...
PPTX
Chengqi zhang graph processing and mining in the era of big data
Machine Status Prediction for Dynamic and Heterogenous Cloud Environment
Web Service QoS Prediction Approach in Mobile Internet Environments
吕潇 星环科技大数据技术探索与应用实践
李战怀 大数据环境下数据存储与管理的研究
2015 07-tuto0-courseoutline
Christian jensen advanced routing in spatial networks using big data
Jeffrey xu yu large graph processing
Calton pu experimental methods on performance in cloud and accuracy in big da...
Ling liu part 02:big graph processing
Ling liu part 01:big graph processing
Wang ke mining revenue-maximizing bundling configuration
Wang ke classification by cut clearance under threshold
2015 07-tuto2-clus type
2015 07-tuto1-phrase mining
2015 07-tuto3-mining hin
2015 07-tuto0-courseoutline
Weiyi meng web data truthfulness analysis
Ke yi small summaries for big data
Gao cong geospatial social media data management and context-aware recommenda...
Chengqi zhang graph processing and mining in the era of big data

Recently uploaded (20)

PPTX
Derivatives of integument scales, beaks, horns,.pptx
PPTX
BIOMOLECULES PPT........................
PDF
CAPERS-LRD-z9:AGas-enshroudedLittleRedDotHostingaBroad-lineActive GalacticNuc...
PPTX
2. Earth - The Living Planet Module 2ELS
PPTX
7. General Toxicologyfor clinical phrmacy.pptx
PDF
HPLC-PPT.docx high performance liquid chromatography
PDF
The scientific heritage No 166 (166) (2025)
DOCX
Viruses (History, structure and composition, classification, Bacteriophage Re...
PPT
The World of Physical Science, • Labs: Safety Simulation, Measurement Practice
PDF
Phytochemical Investigation of Miliusa longipes.pdf
PDF
An interstellar mission to test astrophysical black holes
PPTX
cpcsea ppt.pptxssssssssssssssjjdjdndndddd
PPTX
Classification Systems_TAXONOMY_SCIENCE8.pptx
PDF
IFIT3 RNA-binding activity primores influenza A viruz infection and translati...
PPT
protein biochemistry.ppt for university classes
PDF
AlphaEarth Foundations and the Satellite Embedding dataset
PDF
. Radiology Case Scenariosssssssssssssss
PPTX
Taita Taveta Laboratory Technician Workshop Presentation.pptx
PPTX
neck nodes and dissection types and lymph nodes levels
PPTX
ECG_Course_Presentation د.محمد صقران ppt
Derivatives of integument scales, beaks, horns,.pptx
BIOMOLECULES PPT........................
CAPERS-LRD-z9:AGas-enshroudedLittleRedDotHostingaBroad-lineActive GalacticNuc...
2. Earth - The Living Planet Module 2ELS
7. General Toxicologyfor clinical phrmacy.pptx
HPLC-PPT.docx high performance liquid chromatography
The scientific heritage No 166 (166) (2025)
Viruses (History, structure and composition, classification, Bacteriophage Re...
The World of Physical Science, • Labs: Safety Simulation, Measurement Practice
Phytochemical Investigation of Miliusa longipes.pdf
An interstellar mission to test astrophysical black holes
cpcsea ppt.pptxssssssssssssssjjdjdndndddd
Classification Systems_TAXONOMY_SCIENCE8.pptx
IFIT3 RNA-binding activity primores influenza A viruz infection and translati...
protein biochemistry.ppt for university classes
AlphaEarth Foundations and the Satellite Embedding dataset
. Radiology Case Scenariosssssssssssssss
Taita Taveta Laboratory Technician Workshop Presentation.pptx
neck nodes and dissection types and lymph nodes levels
ECG_Course_Presentation د.محمد صقران ppt

Latent Interest and Topic Mining on User-item Bipartite Networks

  • 1. Latent Interest and Topic Mining on User-item Bipartite Networks Jinliang Xu, Shangguang Wang∗,Sen Su State Key Laboratory of Networking and Switching Technology Beijing University of Posts and Telecommunications, Beijing, China {jlxu,sgwang,susen}@bupt.edu.cn Sathish A.P Kumar Department of Computer Science and Information Systems, Coastal Carolina University South Carolina, USA skumar@coastal.edu Wu Chou Huawei Technologies Co., Ltd Shenzhen, China Wu.Chou@huawei.com Abstract—Latent Factor Model (LFM) is extensively used in dealing with user-item bipartite networks in service recom- mendation systems. To alleviate the limitations of LFM, this papers presents a novel unsupervised learning model, Latent Interest and Topic Mining model (LITM), to automatically mine the latent user interests and item topics from user-item bipartite networks. In particular, we introduce the motivation and objectives of this bipartite network based approach, and detail the model development and optimization process of the proposed LITM. This work not only provides an efficient method for latent user interest and item topic mining, but also highlights a new way to improve the accuracy of service recommendation. Experimental studies are performed and the results validate the LITM’s efficiency in model training, and its ability to provide better service recommendation performance based on user-item bipartite networks are demonstrated. Keywords—service recommendation, user-item bipartite net- work, latent topic and interest, interpretability, and efficiency. I. INTRODUCTION & RELATED WORK Service recommendation systems collect information on the preferences of its users for a set of items to help users identify useful items from a considerably large search space (e.g.,Taobao Mall, Netflix movies). With the rapid develop- ment of e-commerce and mobile internet, new challenges have been posed to researchers who are trying to mine the inner patterns about users, items and relationship between them, such as exploiting the purchasing behaviors in order to identify additional possible matches, i.e. connecting users with items they would otherwise not be able to find. In the study of mining the relationships between users and items, an important and special class of networks is a bipartite network, where nodes can be divided into two disjoint sets, such that no two nodes within the same set are connected [1], [2], [3]. Many systems can be naturally modeled as bipartite networks. For example, users are connected with the movies they have rated (e.g., MovieLens [4]), web-users are connected with the web sites that they have collected in their bookmark sites, and the consumers are connected with the goods that they have purchased from the market. In dealing with the user-item bipartite networks, the tra- ditional Latent Factor Models (LFM) have been shown to be effective [5], [6]. LFM is able to factor the user-item bipartite networks into two smaller matrices, consisting of latent user factor vectors and item factor vectors respectively. However, the output latent factor vectors generated by LFM may contain arbitrary values, including negative numbers, causing low interpretability [7]. As a result, LFM is hard to explain to its users how a specific recommendation is derived. Furthermore, each of the output latent factor vectors are not a probability vector (Note that a probability vector here is a vector with non-negative elements that add up to 1), this property makes LFM difficult for building more complex, realistic models [8]. Similar approaches to recommendation systems include LSA [9], and NMF [10]. On the other hand, LDA [11], [12] can output probabilistic distribution of latent topics when taking user-item bipartite network as input, and it has better interpretability than LFM. However, LDA based models cannot model both users and items [12]. In addition, LDA based models may not be applied directly to mine user interests [2]. For example, Haveliwala et. al [13] computed a set of scores for each item, where each score is related to a topic. However, the topics here can not be automatically mined from the user-item bipartite network. In this paper, by combining the best of LFM and LDA, we propose a novel unsupervised learning model, named Latent Interest and Topic Mining model (LITM) to model both users and items with the better interpretability from user-item bipartite networks. As shown in Fig. 1, LITM is based on the following: 1) for each item there is a probability vector 2016 IEEE International Conference on Services Computing 978-1-5090-2628-9/16 $31.00 © 2016 IEEE DOI 10.1109/SCC.2016.105 778
  • 2. Fig. 1. Concept of latent user interests and item topics in LITM. At the left, it is a typical user-item bipartite network in recommender system, where a user is connected with an item if he has purchased it. The latent user interest and item topic vector are shown at the right side. called latent item topic vector; 2) for each user there is a same length vector called latent user interest vector, in which each position of the latter vector corresponds to the same position of the former vector; and 3) the occurrence probability of each connection is determined by the corresponding user’s latent interest vector and the corresponding item’s latent topic vector. Latent user interest vector and latent item topic vector can generalize the intrinsic patterns underlying the data generation process into smaller number of latent factors. Taking the MovieLens as an example, each movie has a probability dis- tribution over different topics, and each user has a probability distribution of interests over all topics. The remainder of this paper is organized as follows. In Section II we detail the model development and optimization process of the proposed LITM. In Section III, our experimental studies show that the proposed LITM is more efficient in mod- el training against the baseline and it improves the accuracy of service recommendation. Finally, we summarize the findings of this paper in Section IV. II. THE PROPOSED MODEL LITM In this section we provide the details of the model devel- opment and optimization process of LITM. A. Conception and Notations We explicitly describe the concept of latent item topic vectors and latent user interest vectors respectively by the following: • latent item topic vector vm(m = 1, 2, · · · , M): for each item m, we use K-dimensionality probability vector vm to represent its topic probability distribution over K ∈ Z+ topics; • latent user interest vector un(n = 1, 2, · · · N): for each user n, we use the same length probability vector un to represent the probability distribution of user’s interest over the K topics. Each position of un corresponds to the same position in vm. Then we define the occurrence probability of the connection between user n and item m (denoted by un − vm), or the probability that user n select item m, by the following: uT n vm = K k=1 un,kvm,k, (1) where un,k and vm,k are respectively the k−th elements of un and vm. By doing so, we let un,kvm,k be the occurrence probability of un −vm on the k−th topic, then the sum of the occurrence probability over all topics (Eq. 1) can represent the occurrence probability of the connection un − vm. B. LITM Suppose both un(n = 1, 2, · · · N) and vm(m = 1, 2, · · · , M) are known, the occurrence of all connections un − vm in bipartite network G are independent, according to the d-separation criterion in probabilistic graphical models. So we can define the objective function as the occurrence probability of the whole bipartite network G as P(G|U, V ) = N n=1 M m=1 un T vm σm,n , (2) where σm,n indicates whether the connection un − vm exists (σm,n = 1) or not (σm,n = 0), and U = {un|n = 1, 2, · · · , N}, V = {vm|m = 1, 2, · · · , M}. For the ease of optimization, we take the negative logarithm of the objective function, then add two regularized items, and we get the problem to be optimized by the following: minU,V − N n=1 M m=1 σm,n log un T vm + 1 N N n=1 ||un||2 + 1 M M m=1 ||vm||2 s.t. un ≥ 0, ||un||1 = 1, n = 1, 2, · · · , N vm ≥ 0, ||vm||1 = 1, m = 1, 2, · · · , M . (3) Note that we add the latter two 2 norm regularized terms to alleviate the potential overfitting. Analyzing the above problems shows, it shows as follows: 1) This objective function is non-convex (for the un T vm term). Note, however, if we fix U and treat it as constants, then the objective is a convex function of V , and vice versa. We will therefore fix V and optimize U, then fix U and optimize V , thereby obtaining a series of efficiently solvable subproblems. As each update of U and V , it decreases the objective function, from which the process will converge and find a local minimum eventually. 2) If we fix V and optimize U, then un, n = 1, 2, · · · , N can be optimized individually, and vice versa. We take optimizing un as an example, then we get a subproblem like the following: minun − M m=1 σm,n log un T vm + 1 N ||un||2 s.t. un ≥ 0, ||un||1 = 1 , (4) where vm ≥ 0, ||vm||1 = 1, m = 1, 2, · · · , M. Clearly the subproblem (Eq. 4) has no closed-form solution, and cannot use the gradient descent method directly to get the optimal numerical solution for the given constraints un ≥ 0, ||un||1 = 1. However, since this is a convex programming 779
  • 3. problem with linear constraints and according to the convexity- preserving rules, we can apply Reduced Gradient method to search for an optimal solution for the abovementioned subproblems. Algorithm 1: The adaptive Reduced Gradient method Input: σm,n ∈ {0, 1}, vm, m = 1, 2, · · · , M; K ∈ Z+ ; Output: u 1 Initialize u > 0 ∧ ||u||1 = 1, d = 0; 2 while d = 0 do 3 umax = max{u}; 4 uN = uumax //uN contains all but umax; 5 g = − M m=1 σm,n vm uT vm + 2 N un//gradient vector ; 6 gmax = gj|uj =umax // the gradient value at umax; 7 for j = 1; j ≤ K; j + + do 8 if uj ∈ uN then 9 rj = gj − gmax; 10 else 11 loc = j; 12 rj = 0; 13 for j = 1; j ≤ K; j + + do 14 if rj > 0 then 15 dj = −ujrj; 16 else 17 dj = −rj; 18 dloc = −sum{d} //sum{d} is sum value of d; 19 λmax = ∞; 20 for j = 1; j ≤ K; j + + do 21 if dj < 0 then 22 λmax = min λmax, − uj dj 23 search the best step length λbest between [0, λmax]; 24 u = u + λbestd; 25 return u; The Reduced Gradient method uses the equality constraints to eliminate a subset of variables, thereby it reduces the orig- inal problem to a bound-constrained problem in the space of the remaining variables. Algorithm. 1 lists the whole process to get an optimal solution for the subproblem (Eq. 4) using the adaptive Reduced Gradient method. Lines 3-4 select the largest variables as the basic variables. Lines 9-14 ensure the feasibility of the new solution. Lines 15-19 make sure to move in the direction to decrease the objective function. Lines 21-24 calculate the largest possible step length towards the optimizing direction. Based on Algorithm. 1 and the previous analysis, we summarize the basic idea of optimizing the proposed LITM (Eq. 3) in Algorithm. 2, where lines 3-4 means fixing V and optimizing U, line 5-6 means fixing U and optimizing V , and line 2 means repeating the two operations until convergence. This optimization process will lead us to a local minimum of the objective function. Algorithm 2: Basic idea for optimizing LITM Input: σm,n ∈ {0, 1}, n = 1, 2, · · · , N, m = 1, 2, · · · , M; K ∈ Z+ ; Output: un, vm, n = 1, 2, · · · , N, m = 1, 2, · · · , M; 1 Initialize un and vm; 2 while not converged do 3 for n = 1; n ≤ N; n + + do 4 un = minun − M m=1 σm,n log un T vm + 1 N ||un||2 // get the solution by Reduced Gradient method(see Alg.1); 5 for m = 1; m ≤ M; m + + do 6 vm = minvm − N n=1 σm,n log un T vm + 1 M ||vm||2 // get the solution by Reduced Gradient method(see Alg.1); 7 return un, vm, n = 1, 2, · · · , N, m = 1, 2, · · · , M; III. EXPERIMENTAL EVALUATION In this section, we describe a series of experiments that we conducted on benchmarking MovieLens dataset to evaluate our proposed LITM and explore how much it can improve the effectiveness of the service recommendation as well as its efficiency in model training. A. Experimental DataSet & Baseline Model The MovieLens data is from the GroupLens Research1 . It comprises the ratings on 1, 682 items (movies) and 943 users, where each user votes on movies in five discrete ratings 1 ∼ 5. To generate the user-item bipartite network, we consider that the connection between an item and a user exists if and only if the given rating is at least 3. By this way, we obtained a 82, 520 entry (an entry means an actual user-item connection) user-item bipartite network as our experimental dataset. In our experiments, the dataset was randomly partitioned into two parts, where the training set contained 90% of the total entries, and the remaining 10% (8,252 entries) entries constitute the test set. In regards to the baseline model (we set K = 19 currently, and more work is under way), we preliminarily adopted the traditional model LFM [5] that has been described in Section I. We choosed LFM because it has been extensively used in the domain of service recommendation, and similar to our proposed LITM, in LFM each users and items has a latent vector representation. B. Experimental Results & Discussions LFM based recommender systems can provide each user with an ordered queue of all its uncollected items. Similarly our proposed LITM can rank a user’s uncollected items in the order of the occurrence probability of the connections using 1http://www.grouplens.org 780
  • 4. Fig. 2. The predicted position of each entry in the test data ranked in ascending order. Fig. 3. The hitting rate as a function of the length of recommendation list. Fig. 4. Learning curves for LFM and LITM model training. Eq. 1. For a user un, if the connection un − vm is in the test set, we measure the position of vm in the ordered queue. For example, if there are 1000 uncollected items for un, and vm is the 40th from the top, we say the position of vm is the top 40/1000, denoted by rankn,m = 0.04. Since the test entries are actual connections, a good model is expected to give high recommendations to them, indicated by their good rank order positions. Fig. 2 shows the position values of 8, 252 test entries, which are ordered from the top position (rank → 0) to the bottom position (rank → 1). Clearly, the proposed LITM performs better than LFM. To evaluate the proposed LITM in recommender systems from another point of view, we adopted a measure of recom- mending accuracy that depends on the length of recommenda- tion list. The recommendation list for a user un with length L, contains L highest ranking items generated from the model. For each entry un − vm in the test data, if vm is in ui’s recommendation list, we say the entry un −vm is “hit” by the model. The ratio of hit entries is called “hitting rate”. For a given L, the model with a higher hitting rate is better, and vice versa. Clearly, the hitting rate monotonously increases with L, with the upper bound 1 for sufficiently large L. In Fig. 3, we report the hitting rate as a function of L for LFM and LITM, and it’s easy to see that LITM had a better performance than LFM. We also evaluated the efficiency of model training. Both training processes of LFM and LITM will generate a series of solutions in model training, and we plotted the base-10 logarithm of the objective function values of these solutions in Fig. 4. It shows that the proposed LITM converged faster than the traditional LFM, which demonstrates LITM’s efficiency in model training. By changing the learning rate in model training, we obtained the similar results as described above. Despite three distinct metrics being used in the performance evaluation experiments using three distinct metrics, it consis- tently demonstrated the LITM’s efficiency in model training and its ability for better service recommendation performance. IV. CONCLUSION This paper presented a novel unsupervised learning model LITM to automatically mine the latent user interest and item topic distribution from user-item bipartite networks. LITM improves LFM in its low interpretability. In addition, this work not only provides an efficient method for latent user interest and item topic mining, but also highlights a new way to improve accuracy of service recommendation. ACKNOWLEDGMENTS The work presented in this study is supported by NSFC (61472047); NSFC(61571066). REFERENCES [1] T. Hu, H. Xiong, and S. Y. Sung, “Co-preserving patterns in bipartite partitioning for topic identification.,” in SIAM International Conference on Data Mining (SDM), pp. 509–514, SIAM, 2007. [2] X. Tang, M. Zhang, and C. C. Yang, “User interest and topic detec- tion for personalized recommendation,” in Proceedings of the 2012 IEEE/WIC/ACM International Joint Conferences on Web Intelligence and Intelligent Agent Technology (WI-IAT), pp. 442–446, IEEE Com- puter Society, 2012. [3] T. de Paulo Faleiros and A. de Andrade Lopes, “Bipartite graph for topic extraction,” in Proceedings of the 24th International Joint Conference on Artificial Intelligence (IJCAI), pp. 4361–4362, AAAI Press, 2015. [4] F. M. Harper and J. A. Konstan, “The movielens datasets: History and context,” ACM Transactions on Interactive Intelligent Systems, vol. 5, no. 4, p. 19, 2015. [5] R. Bell, Y. Koren, and C. Volinsky, “Modeling relationships at multiple scales to improve accuracy of large recommender systems,” in Proceed- ings of the 13th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 95–104, ACM, 2007. [6] Y. Shen and R. Jin, “Learning personal+ social latent factor model for social recommendation,” in Proceedings of the 18th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 1303–1311, ACM, 2012. [7] G. Friedrich and M. Zanker, “A taxonomy for generating explanations in recommender systems,” AI Magazine, vol. 32, no. 3, pp. 90–98, 2011. [8] S. R. Eddy, “A probabilistic model of local sequence alignment that sim- plifies statistical significance estimation,” PLOS Computational Biology, vol. 4, no. 5, p. e1000069, 2008. [9] T. Hofmann, “Probabilistic latent semantic indexing,” in Proceedings of the 22nd Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 50–57, ACM, 1999. [10] D. D. Lee and H. S. Seung, “Algorithms for non-negative matrix factorization,” in Proceedings of the Annual Conference on Neural Information Processing Systems (NIPS), pp. 556–562, 2001. [11] D. M. Blei, A. Y. Ng, and M. I. Jordan, “Latent dirichlet allocation,” Journal of Machine Learning Research, vol. 3, pp. 993–1022, 2003. [12] J. Tang, R. Jin, and J. Zhang, “A topic modeling approach and its integration into the random walk framework for academic search,” in Proceedings of the 8th IEEE International Conference on Data Mining (ICDM), pp. 1055–1060, IEEE, 2008. [13] T. H. Haveliwala, “Topic-sensitive pagerank: A context-sensitive ranking algorithm for web search,” IEEE Transactions on Knowledge and Data Engineering, vol. 15, no. 4, pp. 784–796, 2003. 781