SlideShare a Scribd company logo
Department of Business
Information Systems II
Using Siamese Graph Neural Networks for
Similarity-Based Retrieval in Process-Oriented
Case-Based Reasoning
Maximilian Hoffmann, Lukas Malburg, Patrick Klein,
and Ralph Bergmann
Department of Business Information Systems II
University of Trier, Germany
www.wi2.uni-trier.de
hoffmannm@uni-trier.de
Funded by (BE 1373/3-3 and 375342983)
Department of Business
Information Systems II
Outline
• Motivation
• Related Work
• Approach for Learning Similarities of Workflow Graphs via Siamese
Graph Neural Network (GNNs)
• Experimental Evaluation
• Conclusion and Future Work
- 2 -
Department of Business
Information Systems II
Motivation
• Semantic graphs as cases in Process-Oriented CBR:
– Similarity assessment: kind of inexact sub-graph matching
– Similarity considers structure of nodes and edges and their semantic annotations
– High complexity leads to long retrieval times
• Two-phase MAC/FAC retrieval
– MAC: similarity measure for a fast pre-selection
– FAC: graph matching is applied
– Existing automatically-learned POCBR MAC/FAC approaches use embeddings …
• Advantage: Usage of fast and simple vector similarity measures
– … but complex semantics are handled insufficiently
• No consideration of graph structure and semantic annotations
• Decreased utility for rather complex domains
• Advantages of embedding representation should be adapted for other
approaches
- 3 -
Goal: Similarity assessment of complex graphs via siamese GNNs
Department of Business
Information Systems II
Related Work
1. Embedding-based approach by Klein et al. (2019)
• Automatically-learned low-dimensional embedding vectors for graph nodes
• Similarity of two graphs given by aggregating node vectors and applying a
vector similarity measure
• Very fast method, but lacks proper integration of graph structure and semantic
annotations into embedding vectors
2. Cluster-based approach by Müller and Bergmann (2014)
• based on the cluster structure of the case base
• finding clusters that are similar to the query
• no additional modeling effort
• not reaching performance of the feature-based approach
3. Domain specific feature-based case representation by Bergmann and
Stromer (2013)
• simplified version of native representation
• similarity computation only uses feature representation and not the graph
• requires additional modeling effort (manually modeled)
- 4 -
Department of Business
Information Systems II
Graph Similarity Assessment with
Siamese GNNs
• Challenges:
1. Transforming all relevant information of semantic graphs into a
format, interpretable by neural networks
2. Finding a siamese GNN that is capable of producing meaningful
graph similarities and short retrieval times
• Contributions:
– Encoding method of semantic graphs
• Supports complex semantic descriptions, including nested entries
• Supports the complete graph structure with all nodes and edges
– Two siamese GNN architectures by Li et al. (2019)
• Different levels of computational complexity
• Modified for retrieving semantic graphs
- 5 -
Department of Business
Information Systems II
Encoding Semantic Graphs for
Similarity Learning
• Composition of node and edge encodings:
– Encoding of node and edge type
– Encoding of semantic annotations and ProCAKE data types such
as key-value pairs, lists, numerics, strings, taxonomies or sets.
• Sequence of vectors can be transformed to a matrix
• Matrix is processable by neural networks
- 6 -
Department of Business
Information Systems II
Graph Embedding Model (GEM)
- 7 -
• Encoder:
− Transforms raw encoded data
into node and edge embeddings
− Usage of Feed-Forward- and
Recurrent Neural Networks
• Propagation Layer:
− Propagates embeddings of con-
nected nodes in both graphs
− Captures information on the
neighborhood of individual nodes
• Aggregator:
− Merges final node embeddings
to a whole-graph embedding
− Pairwise vector similarity
Complexity O(n)
Department of Business
Information Systems II
Graph Matching Network (GMN)
- 8 -
• Encoder:
− Transforms raw encoded data
into node and edge embeddings
− Usage of Feed-Forward- and
Recurrent Neural Networks
• Propagation Layer:
− Same goal as propagation layer
in GEM but different scope
− Propagates information across
graphs via attention mechanism
• Aggregator:
− Merges final node embeddings
to a whole-graph embedding
− Pairwise similarity using a Feed-
Forward Neural Network
Complexity O(n²)
Department of Business
Information Systems II
Graph Retrieval using GEM or GMN
• Retrieval utilizes one of the neural networks (GEM/GMN):
– Input data: Encoded semantic graphs
– Output data: pairwise graph similarities
– Prediction of the similarity of each case with the query
– K-most similar cases are determined
– Both neural networks are also applicable in a MAC/FAC setup
• Implementation uses ProCAKE and Tensorflow
- 9 -
(https://procake.uni-trier.de)
Department of Business
Information Systems II
Experimental Evaluation
• Setup:
– Comparison of GEM and GMN to the embedding-based retriever (EBR)
[1] and the feature-based retriever (FBR) [2] in two experiments
1. MAC/FAC retrieval with different filter sizes and values of k (similar to [1])
2. Approximation of the A-Star retriever (A*R) by Bergmann and Gil (2014) [3]
– Experiments for two workflow domains:
• CB-I: Simple cooking recipes (680 training and 120 testing cases)
• CB-II: Complex data mining workflows (529 training and 80 testing cases)
– Examination of retrieval quality and performance
• Hypotheses:
– H1: Using GEM and GMN as a MAC retriever of a MAC/FAC retrieval
leads to better retrieval results than using EBR as MAC retriever.
– H2: The GMN retriever is able to approximate the ground-truth graph
similarities better than A*R, using parameter settings such that the
retrieval time of both retrievers is comparable.
- 10 -
Department of Business
Information Systems II
Experimental Results (MAC/FAC)
• CB-I:
– FBR has the best qualities and GEM and EBR are the fastest retrievers
– GEM and GMN are not able to outperform EBR
• CB-II:
– FBR still performs very well, GEM and EBR are still the fastest retrievers
– GEM and GMN now outperform EBR
• H1 is partly confirmed: Rejected for CB-I and accepted for CB-II
- 11 -
Department of Business
Information Systems II
Experimental Results (A-Star
Approximation)
• CB-I: GMN has lowest MAE
and A*R highest correctness
• CB-II: GMN has lowest MAE
and highest correctness
• H2: Clearly accepted for CB-II
and partly accepted for CB-I
- 12 -
Department of Business
Information Systems II
Conclusion and Future Work
Conclusion
• GEM and GMN show high potential in graph retrieval scenarios, especially in
more complex domains
– GEM as a MAC measure outperforms other machine-learned approaches
– GMN as a FAC measure outperforms an A-Star-based graph matching measure
Future Work
• Optimization of the presented neural networks
– Usage of a differentiable ranking loss function for GEM
– Optimized encoding scheme for more data types and other graph structures
• Evaluation on other domains, e.g., argument graphs, and other types of
complex graph similarity measures
• Investigation of suitable methods for explaining the results of the neural
networks in the context of Explainable Artificial Intelligence (e.g., see XAI
workshop of ICCBR 2019)
- 13 -
Department of Business
Information Systems II - 14 -
Thank you for your
attention!
Department of Business
Information Systems II
Department of Business
Information Systems II
References
[1] Klein, P., Malburg, L., Bergmann, R.: Learning Workflow Embeddings to Improve
the Performance of Similarity-Based Retrieval for Process-Oriented Case-Based
Reasoning. In: Case-Based Reasoning Research and Development: 27th Inter-
Conference, ICCBR 2019, Germany, pp. 188–203. Springer. (2019)
[2] Bergmann, R., Stromer, A.: MAC/FAC Retrieval of Semantic Workflows. In:
Boonthum-Denecke, C., Youngblood, G.M. (eds.) Proceedings of the Twenty-
Sixth International Florida Artificial Intelligence Research Society Conference,
FLAIRS 2013. AAAI Press (2013)
[3] Bergmann, R., Gil, Y.: Similarity assessment and efficient retrieval of semantic
workflows. Information Systems 40, pp. 115–127 (2014)
[4] Li, Y., Gu, C., Dullien, T., Vinyals, O., Kohli, P.: Graph Matching Networks for
Learning the Similarity of Graph Structured Objects. In: Chaudhuri, K.,
Salakhutdinov, R. (eds.) Proc. of the 36th Int. Conf. on Machine Learning, ICML
2019, USA. Proc. of Machine Learning Research, vol. 97, pp. 3835–3845. PMLR
- 16 -

More Related Content

PDF
Informed Machine Learning for Improved Similarity Assessment in Process-Orien...
PDF
hetero_pim
PDF
A METHODOLOGY FOR IMPROVEMENT OF ROBA MULTIPLIER FOR ELECTRONIC APPLICATIONS
PDF
deepgbm_kdd2019__CR_.pdf
PDF
V3I8-0460
DOCX
An optimized modified booth recoder for efficient design of the add multiply ...
PDF
Graph Signal Processing for Machine Learning A Review and New Perspectives - ...
PDF
Towards Systematic Parallel Programming over MapReduce
Informed Machine Learning for Improved Similarity Assessment in Process-Orien...
hetero_pim
A METHODOLOGY FOR IMPROVEMENT OF ROBA MULTIPLIER FOR ELECTRONIC APPLICATIONS
deepgbm_kdd2019__CR_.pdf
V3I8-0460
An optimized modified booth recoder for efficient design of the add multiply ...
Graph Signal Processing for Machine Learning A Review and New Perspectives - ...
Towards Systematic Parallel Programming over MapReduce

What's hot (18)

PDF
Graph Signal Processing for Machine Learning A Review and New Perspectives - ...
PDF
Efficient implementation of bit parallel finite
PDF
Efficient implementation of bit parallel finite field multipliers
PDF
Learning Graph Representation for Data-Efficiency RL
PDF
Interior Dual Optimization Software Engineering with Applications in BCS Elec...
PPTX
Locally densest subgraph discovery
PDF
Parallel Machine Learning
PPTX
A Graph Summarization: A Survey | Summarizing and understanding large graphs
PDF
HOCSA: AN EFFICIENT DOWNLINK BURST ALLOCATION ALGORITHM TO ACHIEVE HIGH FRAME...
PDF
Recurrent Neural Networks for Recommendations and Personalization with Nick P...
PDF
Applying Neural Networks and Analogous Estimating to Determine the Project Bu...
PDF
Graph Signal Processing for Machine Learning A Review and New Perspectives - ...
PDF
Transformation and dynamic visualization of images from computer through an F...
PDF
A mathematical model and a heuristic memory allocation problem
PPT
An Adaptive Load Balancing Middleware for Distributed Simulation
PPTX
Distributed graph summarization
PDF
Integration of a Predictive, Continuous Time Neural Network into Securities M...
PDF
Jf3515881595
Graph Signal Processing for Machine Learning A Review and New Perspectives - ...
Efficient implementation of bit parallel finite
Efficient implementation of bit parallel finite field multipliers
Learning Graph Representation for Data-Efficiency RL
Interior Dual Optimization Software Engineering with Applications in BCS Elec...
Locally densest subgraph discovery
Parallel Machine Learning
A Graph Summarization: A Survey | Summarizing and understanding large graphs
HOCSA: AN EFFICIENT DOWNLINK BURST ALLOCATION ALGORITHM TO ACHIEVE HIGH FRAME...
Recurrent Neural Networks for Recommendations and Personalization with Nick P...
Applying Neural Networks and Analogous Estimating to Determine the Project Bu...
Graph Signal Processing for Machine Learning A Review and New Perspectives - ...
Transformation and dynamic visualization of images from computer through an F...
A mathematical model and a heuristic memory allocation problem
An Adaptive Load Balancing Middleware for Distributed Simulation
Distributed graph summarization
Integration of a Predictive, Continuous Time Neural Network into Securities M...
Jf3515881595
Ad

Similar to Using Siamese Graph Neural Networks for Similarity-Based Retrieval in Process-Oriented Case-Based Reasoning (20)

PPTX
RNNs for Recommendations and Personalization
PDF
PRETZEL: Opening the Black Box of Machine Learning Prediction Serving Systems
PPTX
NS-CUK Seminar: S.T.Nguyen, Review on "Do We Really Need Complicated Model Ar...
PDF
IRJET- Recommendation System based on Graph Database Techniques
PPTX
GRID COMPUTING
PDF
STOCK MARKET PREDICTION USING NEURAL NETWORKS
PDF
Supervised embedding techniques in search ranking system
PPTX
bundle__block_adjustmentpropert_ppt.pptx
PDF
Parallel KNN for Big Data using Adaptive Indexing
PPTX
MLIntro_ADA.pptx
PDF
D04573033
PPTX
Scalable Data Analytics: Technologies and Methods
PDF
Memory Efficient Graph Convolutional Network based Distributed Link Prediction
PPTX
Simple Presentation in Pink Lilac Pastel Blobs Basic Style.pptx
PDF
Deep learning for 3 d point clouds presentation
PPTX
Deep Learning for Recommender Systems
PDF
IRJET- Efficient Auto Annotation for Tag and Image based Searching Over Large...
PDF
[White paper] Maintain-Accurate-Network-Diagrams
PDF
Measuring method complexity of the case management modeling and notation (CMMN)
PPTX
Optimal Chain Matrix Multiplication Big Data Perspective
RNNs for Recommendations and Personalization
PRETZEL: Opening the Black Box of Machine Learning Prediction Serving Systems
NS-CUK Seminar: S.T.Nguyen, Review on "Do We Really Need Complicated Model Ar...
IRJET- Recommendation System based on Graph Database Techniques
GRID COMPUTING
STOCK MARKET PREDICTION USING NEURAL NETWORKS
Supervised embedding techniques in search ranking system
bundle__block_adjustmentpropert_ppt.pptx
Parallel KNN for Big Data using Adaptive Indexing
MLIntro_ADA.pptx
D04573033
Scalable Data Analytics: Technologies and Methods
Memory Efficient Graph Convolutional Network based Distributed Link Prediction
Simple Presentation in Pink Lilac Pastel Blobs Basic Style.pptx
Deep learning for 3 d point clouds presentation
Deep Learning for Recommender Systems
IRJET- Efficient Auto Annotation for Tag and Image based Searching Over Large...
[White paper] Maintain-Accurate-Network-Diagrams
Measuring method complexity of the case management modeling and notation (CMMN)
Optimal Chain Matrix Multiplication Big Data Perspective
Ad

Recently uploaded (20)

PPTX
2. Earth - The Living Planet Module 2ELS
PDF
bbec55_b34400a7914c42429908233dbd381773.pdf
PPTX
Classification Systems_TAXONOMY_SCIENCE8.pptx
PDF
Placing the Near-Earth Object Impact Probability in Context
PPT
POSITIONING IN OPERATION THEATRE ROOM.ppt
PPTX
INTRODUCTION TO EVS | Concept of sustainability
PPTX
ECG_Course_Presentation د.محمد صقران ppt
PDF
The scientific heritage No 166 (166) (2025)
PDF
Unveiling a 36 billion solar mass black hole at the centre of the Cosmic Hors...
PDF
AlphaEarth Foundations and the Satellite Embedding dataset
PPTX
ANEMIA WITH LEUKOPENIA MDS 07_25.pptx htggtftgt fredrctvg
PPTX
Protein & Amino Acid Structures Levels of protein structure (primary, seconda...
PDF
Phytochemical Investigation of Miliusa longipes.pdf
PPTX
TOTAL hIP ARTHROPLASTY Presentation.pptx
PPTX
2. Earth - The Living Planet earth and life
PPTX
Comparative Structure of Integument in Vertebrates.pptx
PDF
IFIT3 RNA-binding activity primores influenza A viruz infection and translati...
PDF
An interstellar mission to test astrophysical black holes
PPTX
Microbiology with diagram medical studies .pptx
PDF
HPLC-PPT.docx high performance liquid chromatography
2. Earth - The Living Planet Module 2ELS
bbec55_b34400a7914c42429908233dbd381773.pdf
Classification Systems_TAXONOMY_SCIENCE8.pptx
Placing the Near-Earth Object Impact Probability in Context
POSITIONING IN OPERATION THEATRE ROOM.ppt
INTRODUCTION TO EVS | Concept of sustainability
ECG_Course_Presentation د.محمد صقران ppt
The scientific heritage No 166 (166) (2025)
Unveiling a 36 billion solar mass black hole at the centre of the Cosmic Hors...
AlphaEarth Foundations and the Satellite Embedding dataset
ANEMIA WITH LEUKOPENIA MDS 07_25.pptx htggtftgt fredrctvg
Protein & Amino Acid Structures Levels of protein structure (primary, seconda...
Phytochemical Investigation of Miliusa longipes.pdf
TOTAL hIP ARTHROPLASTY Presentation.pptx
2. Earth - The Living Planet earth and life
Comparative Structure of Integument in Vertebrates.pptx
IFIT3 RNA-binding activity primores influenza A viruz infection and translati...
An interstellar mission to test astrophysical black holes
Microbiology with diagram medical studies .pptx
HPLC-PPT.docx high performance liquid chromatography

Using Siamese Graph Neural Networks for Similarity-Based Retrieval in Process-Oriented Case-Based Reasoning

  • 1. Department of Business Information Systems II Using Siamese Graph Neural Networks for Similarity-Based Retrieval in Process-Oriented Case-Based Reasoning Maximilian Hoffmann, Lukas Malburg, Patrick Klein, and Ralph Bergmann Department of Business Information Systems II University of Trier, Germany www.wi2.uni-trier.de hoffmannm@uni-trier.de Funded by (BE 1373/3-3 and 375342983)
  • 2. Department of Business Information Systems II Outline • Motivation • Related Work • Approach for Learning Similarities of Workflow Graphs via Siamese Graph Neural Network (GNNs) • Experimental Evaluation • Conclusion and Future Work - 2 -
  • 3. Department of Business Information Systems II Motivation • Semantic graphs as cases in Process-Oriented CBR: – Similarity assessment: kind of inexact sub-graph matching – Similarity considers structure of nodes and edges and their semantic annotations – High complexity leads to long retrieval times • Two-phase MAC/FAC retrieval – MAC: similarity measure for a fast pre-selection – FAC: graph matching is applied – Existing automatically-learned POCBR MAC/FAC approaches use embeddings … • Advantage: Usage of fast and simple vector similarity measures – … but complex semantics are handled insufficiently • No consideration of graph structure and semantic annotations • Decreased utility for rather complex domains • Advantages of embedding representation should be adapted for other approaches - 3 - Goal: Similarity assessment of complex graphs via siamese GNNs
  • 4. Department of Business Information Systems II Related Work 1. Embedding-based approach by Klein et al. (2019) • Automatically-learned low-dimensional embedding vectors for graph nodes • Similarity of two graphs given by aggregating node vectors and applying a vector similarity measure • Very fast method, but lacks proper integration of graph structure and semantic annotations into embedding vectors 2. Cluster-based approach by Müller and Bergmann (2014) • based on the cluster structure of the case base • finding clusters that are similar to the query • no additional modeling effort • not reaching performance of the feature-based approach 3. Domain specific feature-based case representation by Bergmann and Stromer (2013) • simplified version of native representation • similarity computation only uses feature representation and not the graph • requires additional modeling effort (manually modeled) - 4 -
  • 5. Department of Business Information Systems II Graph Similarity Assessment with Siamese GNNs • Challenges: 1. Transforming all relevant information of semantic graphs into a format, interpretable by neural networks 2. Finding a siamese GNN that is capable of producing meaningful graph similarities and short retrieval times • Contributions: – Encoding method of semantic graphs • Supports complex semantic descriptions, including nested entries • Supports the complete graph structure with all nodes and edges – Two siamese GNN architectures by Li et al. (2019) • Different levels of computational complexity • Modified for retrieving semantic graphs - 5 -
  • 6. Department of Business Information Systems II Encoding Semantic Graphs for Similarity Learning • Composition of node and edge encodings: – Encoding of node and edge type – Encoding of semantic annotations and ProCAKE data types such as key-value pairs, lists, numerics, strings, taxonomies or sets. • Sequence of vectors can be transformed to a matrix • Matrix is processable by neural networks - 6 -
  • 7. Department of Business Information Systems II Graph Embedding Model (GEM) - 7 - • Encoder: − Transforms raw encoded data into node and edge embeddings − Usage of Feed-Forward- and Recurrent Neural Networks • Propagation Layer: − Propagates embeddings of con- nected nodes in both graphs − Captures information on the neighborhood of individual nodes • Aggregator: − Merges final node embeddings to a whole-graph embedding − Pairwise vector similarity Complexity O(n)
  • 8. Department of Business Information Systems II Graph Matching Network (GMN) - 8 - • Encoder: − Transforms raw encoded data into node and edge embeddings − Usage of Feed-Forward- and Recurrent Neural Networks • Propagation Layer: − Same goal as propagation layer in GEM but different scope − Propagates information across graphs via attention mechanism • Aggregator: − Merges final node embeddings to a whole-graph embedding − Pairwise similarity using a Feed- Forward Neural Network Complexity O(n²)
  • 9. Department of Business Information Systems II Graph Retrieval using GEM or GMN • Retrieval utilizes one of the neural networks (GEM/GMN): – Input data: Encoded semantic graphs – Output data: pairwise graph similarities – Prediction of the similarity of each case with the query – K-most similar cases are determined – Both neural networks are also applicable in a MAC/FAC setup • Implementation uses ProCAKE and Tensorflow - 9 - (https://procake.uni-trier.de)
  • 10. Department of Business Information Systems II Experimental Evaluation • Setup: – Comparison of GEM and GMN to the embedding-based retriever (EBR) [1] and the feature-based retriever (FBR) [2] in two experiments 1. MAC/FAC retrieval with different filter sizes and values of k (similar to [1]) 2. Approximation of the A-Star retriever (A*R) by Bergmann and Gil (2014) [3] – Experiments for two workflow domains: • CB-I: Simple cooking recipes (680 training and 120 testing cases) • CB-II: Complex data mining workflows (529 training and 80 testing cases) – Examination of retrieval quality and performance • Hypotheses: – H1: Using GEM and GMN as a MAC retriever of a MAC/FAC retrieval leads to better retrieval results than using EBR as MAC retriever. – H2: The GMN retriever is able to approximate the ground-truth graph similarities better than A*R, using parameter settings such that the retrieval time of both retrievers is comparable. - 10 -
  • 11. Department of Business Information Systems II Experimental Results (MAC/FAC) • CB-I: – FBR has the best qualities and GEM and EBR are the fastest retrievers – GEM and GMN are not able to outperform EBR • CB-II: – FBR still performs very well, GEM and EBR are still the fastest retrievers – GEM and GMN now outperform EBR • H1 is partly confirmed: Rejected for CB-I and accepted for CB-II - 11 -
  • 12. Department of Business Information Systems II Experimental Results (A-Star Approximation) • CB-I: GMN has lowest MAE and A*R highest correctness • CB-II: GMN has lowest MAE and highest correctness • H2: Clearly accepted for CB-II and partly accepted for CB-I - 12 -
  • 13. Department of Business Information Systems II Conclusion and Future Work Conclusion • GEM and GMN show high potential in graph retrieval scenarios, especially in more complex domains – GEM as a MAC measure outperforms other machine-learned approaches – GMN as a FAC measure outperforms an A-Star-based graph matching measure Future Work • Optimization of the presented neural networks – Usage of a differentiable ranking loss function for GEM – Optimized encoding scheme for more data types and other graph structures • Evaluation on other domains, e.g., argument graphs, and other types of complex graph similarity measures • Investigation of suitable methods for explaining the results of the neural networks in the context of Explainable Artificial Intelligence (e.g., see XAI workshop of ICCBR 2019) - 13 -
  • 14. Department of Business Information Systems II - 14 - Thank you for your attention!
  • 16. Department of Business Information Systems II References [1] Klein, P., Malburg, L., Bergmann, R.: Learning Workflow Embeddings to Improve the Performance of Similarity-Based Retrieval for Process-Oriented Case-Based Reasoning. In: Case-Based Reasoning Research and Development: 27th Inter- Conference, ICCBR 2019, Germany, pp. 188–203. Springer. (2019) [2] Bergmann, R., Stromer, A.: MAC/FAC Retrieval of Semantic Workflows. In: Boonthum-Denecke, C., Youngblood, G.M. (eds.) Proceedings of the Twenty- Sixth International Florida Artificial Intelligence Research Society Conference, FLAIRS 2013. AAAI Press (2013) [3] Bergmann, R., Gil, Y.: Similarity assessment and efficient retrieval of semantic workflows. Information Systems 40, pp. 115–127 (2014) [4] Li, Y., Gu, C., Dullien, T., Vinyals, O., Kohli, P.: Graph Matching Networks for Learning the Similarity of Graph Structured Objects. In: Chaudhuri, K., Salakhutdinov, R. (eds.) Proc. of the 36th Int. Conf. on Machine Learning, ICML 2019, USA. Proc. of Machine Learning Research, vol. 97, pp. 3835–3845. PMLR - 16 -