SlideShare a Scribd company logo
Algorithms
Nikita Kapil
Kohonen’s self-organizing
map (K-SOM)
Artificial Neural Network - Used to reduce dimensionality
and effectively represent n features of a given dataset into 2
features (n-d graph to 2-d graph)
Kohonen’s Self-Organizing Map (SOM)
• Each node’s weights are initialized
• One vector (instance) from the dataset is taken
• Every node is examined to calculate which one’s weights are most like the input
vector. The winning node is commonly known as the Best Matching
Unit (BMU), given by
• Then the neighborhood of the BMU is calculated. The amount of neighbors
decreases over time.
• The winning weight is rewarded with becoming more like the sample vector.
• The next vector is then taken from the dataset. These steps repeat till all data
points are calibrated.
Algorithm explanations
K-Means Clustering (K-MC)
Instance-based lazy learning - Used to cluster data
such that similar data points are represented together
and dissimilar data points are further away.
K-means clustering (k-mc)
• For k number of clusters, choose k random data points as centroids of
the clusters.
• Now choose one data point after another (other than the centroid
points) and calculate distance (preferably Euclidean), given by
• Cluster the data point with the nearest centroid and recalculate the
centroid taking mean of coordinates
• Repeat for other data points
Algorithm explanations
Logistic Regression (LR)
Instance based quick learning - Used to map dependent
variables to an independent variable and separate the
positive and negative instances.
Logistic regression (LR)
• Take data point
• Find class probability using sigmoid function as
follows:
• Then give the data a threshold such that if the value
of g(z) is above that threshold then it is considered
positive (1) otherwise it is considered negative (0)
Algorithm explanations
Support Vector machine
(SVM)
Instance-based quick learning - Used to make a
hyperplane that separates the positive and negative
samples with highest amount of margin
Support Vector Machine (SVM)
• The data points are first plotted in an n-dimensional space.
• Many different lines, planes or hyperplanes (based on number of
dimensions) are plotted to separate the two cases
• The line/plane/hyperplane is chosen which has a maximum
amount of margin between the positive and negative data points.
This can be represented as
• The above formula can be used to identify positive classes
Algorithm explanations
C4.5 decision tree (DT)
Decision Tree based learning - Used for binary classification
(two outcomes) in a question-answer tree format, with most
relevant questions at the top and least relevant questions at
the bottom
C4.5 Decision tree (DT)
• Find the feature of data that lends most information towards the
outcome by using a method (generally least-error, information-gain,
or gini coefficient)
• Place that feature as the root node, and draw out branches, one for
each of the values of that feature, and then assign child node as
follows:
• If that value of the node is decisive (i.e., if that value gives a
decisive outcome) put the outcome as the leaf node
• Else if the value is indecisive, repeat the above steps to calculate
the sub-tree with the data used being the sub-dataset where that
value exists
Algorithm explanations
Random forest (RF)
Decision Tree based learning - An ensemble model
that creates a lot of random trees and takes majority
vote of their decisions
Random forest (RF)
• Create n decision trees randomly based on the
features in the dataset, such that one tree may be a
subset of another but no tree is the exact same
• Obtain the result of all the decision trees and take
majority vote to get the result
Algorithm explanations
Gradient boosting decision
tree (GBdt)
Decision Tree based greedy learning - An advanced
decision tree that assigns weights to questions and
calibrates weight to get a maximum accuracy tree.
Gradient boosting decision tree (gbdt)
• Build a decision tree similarly to c4.5 decision trees, but by assigning random
initial weights to the given questions and sorting according to descending order
of weights.
• Assign a learning rate which defines how quick the tree will change.
• Predict values for a data instance as (Prediction = Average value + learning
rate x weighted increment)
• Find the difference between the results (Difference = correct result - predicted
value)
• If the result is wrong, then adjust the weights for the next tree such that the
overall result is closer to the result of that instance.
• Take the summative results of all the given decision trees and classify
accordingly
Algorithm explanations
K-nearest neighbors (knn)
Instance based lazy learning - Simple, lazy learning
algorithm which classifies the given data points according to
the classes the nearest points.
K-nearest neighbors (knn)
• Plot all the known classified data points into an n-dimensional space
• Consider a point whose dimensional coordinates are known (test point) but
whose class is unknown
• Compute distance as
• Find the nearest 1 neighbor and consider its class
• Continue considering more and more points such that the decision considers a
lot of positive and negative class points with as much confidence as possible.
• Use the most confident class as the class of the test point
Algorithm explanations

More Related Content

PPTX
KNN Algorithm using C++
PPTX
K Nearest Neighbor Algorithm
PDF
K - Nearest neighbor ( KNN )
PPTX
Unsupervised Learning
PDF
Training machine learning knn 2017
PPTX
Unsupervised Learning
PDF
K means Clustering
KNN Algorithm using C++
K Nearest Neighbor Algorithm
K - Nearest neighbor ( KNN )
Unsupervised Learning
Training machine learning knn 2017
Unsupervised Learning
K means Clustering

What's hot (20)

PDF
Lecture 8: Decision Trees & k-Nearest Neighbors
PPTX
Large Scale Data Clustering: an overview
PPT
3.3 hierarchical methods
PPT
Cure, Clustering Algorithm
PDF
Birch
PPTX
Clustering on database systems rkm
PPTX
K-means Clustering
PPTX
05 k-means clustering
PPT
3.2 partitioning methods
PPT
Intro to MATLAB and K-mean algorithm
PPT
Enhance The K Means Algorithm On Spatial Dataset
PPTX
Knn Algorithm presentation
PPTX
Instance based learning
PPTX
ML basic & clustering
PDF
Mean shift and Hierarchical clustering
PPTX
Nearest neighbors
PPT
PDF
K means and dbscan
PPTX
K means clustring @jax
Lecture 8: Decision Trees & k-Nearest Neighbors
Large Scale Data Clustering: an overview
3.3 hierarchical methods
Cure, Clustering Algorithm
Birch
Clustering on database systems rkm
K-means Clustering
05 k-means clustering
3.2 partitioning methods
Intro to MATLAB and K-mean algorithm
Enhance The K Means Algorithm On Spatial Dataset
Knn Algorithm presentation
Instance based learning
ML basic & clustering
Mean shift and Hierarchical clustering
Nearest neighbors
K means and dbscan
K means clustring @jax
Ad

Similar to Algorithm explanations (20)

PPTX
demo lecture for foundation class for btech
PPTX
Supervised Learning Algorithm Slide.pptx
DOCX
Performance analysis of machine learning algorithms on self localization system1
PPTX
Introduction to data visualization tools like Tableau and Power BI and Excel
PPTX
Macine learning algorithms - K means, KNN
PDF
Machine Learning Algorithms Introduction.pdf
PPTX
A_short_brief_on_Machine-Learning's and there types
PPTX
5. Machine Learning.pptx
PPTX
Machine Learning techniques used in AI.
PPTX
How Machine Learning Helps Organizations to Work More Efficiently?
PPTX
Deep learning from mashine learning AI..
PPTX
Reuqired ppt for machine learning algirthms and part
PPTX
Intro to machine learning
PPTX
DECESION TREE and -SVM-NAIVEs bayes-BAYS.pptx
DOCX
introduction to machine learning unit iv
PPTX
ML SFCSE.pptx
PPTX
AI Algorithms
PPTX
Machine Learning: Transforming Data into Insights
PPTX
Different Algorithms used in classification [Auto-saved].pptx
PPT
Introduction to Machine Learning Aristotelis Tsirigos
demo lecture for foundation class for btech
Supervised Learning Algorithm Slide.pptx
Performance analysis of machine learning algorithms on self localization system1
Introduction to data visualization tools like Tableau and Power BI and Excel
Macine learning algorithms - K means, KNN
Machine Learning Algorithms Introduction.pdf
A_short_brief_on_Machine-Learning's and there types
5. Machine Learning.pptx
Machine Learning techniques used in AI.
How Machine Learning Helps Organizations to Work More Efficiently?
Deep learning from mashine learning AI..
Reuqired ppt for machine learning algirthms and part
Intro to machine learning
DECESION TREE and -SVM-NAIVEs bayes-BAYS.pptx
introduction to machine learning unit iv
ML SFCSE.pptx
AI Algorithms
Machine Learning: Transforming Data into Insights
Different Algorithms used in classification [Auto-saved].pptx
Introduction to Machine Learning Aristotelis Tsirigos
Ad

Recently uploaded (20)

PPTX
additive manufacturing of ss316l using mig welding
PDF
Arduino robotics embedded978-1-4302-3184-4.pdf
PDF
BMEC211 - INTRODUCTION TO MECHATRONICS-1.pdf
PPTX
UNIT-1 - COAL BASED THERMAL POWER PLANTS
PPTX
CH1 Production IntroductoryConcepts.pptx
DOCX
573137875-Attendance-Management-System-original
PDF
keyrequirementskkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkk
PPTX
bas. eng. economics group 4 presentation 1.pptx
PPTX
Sustainable Sites - Green Building Construction
PDF
Evaluating the Democratization of the Turkish Armed Forces from a Normative P...
PPTX
FINAL REVIEW FOR COPD DIANOSIS FOR PULMONARY DISEASE.pptx
PPT
Mechanical Engineering MATERIALS Selection
PPTX
Lesson 3_Tessellation.pptx finite Mathematics
PPTX
Strings in CPP - Strings in C++ are sequences of characters used to store and...
PPTX
Construction Project Organization Group 2.pptx
PDF
The CXO Playbook 2025 – Future-Ready Strategies for C-Suite Leaders Cerebrai...
PDF
Operating System & Kernel Study Guide-1 - converted.pdf
PDF
Model Code of Practice - Construction Work - 21102022 .pdf
PPTX
IOT PPTs Week 10 Lecture Material.pptx of NPTEL Smart Cities contd
PDF
Well-logging-methods_new................
additive manufacturing of ss316l using mig welding
Arduino robotics embedded978-1-4302-3184-4.pdf
BMEC211 - INTRODUCTION TO MECHATRONICS-1.pdf
UNIT-1 - COAL BASED THERMAL POWER PLANTS
CH1 Production IntroductoryConcepts.pptx
573137875-Attendance-Management-System-original
keyrequirementskkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkk
bas. eng. economics group 4 presentation 1.pptx
Sustainable Sites - Green Building Construction
Evaluating the Democratization of the Turkish Armed Forces from a Normative P...
FINAL REVIEW FOR COPD DIANOSIS FOR PULMONARY DISEASE.pptx
Mechanical Engineering MATERIALS Selection
Lesson 3_Tessellation.pptx finite Mathematics
Strings in CPP - Strings in C++ are sequences of characters used to store and...
Construction Project Organization Group 2.pptx
The CXO Playbook 2025 – Future-Ready Strategies for C-Suite Leaders Cerebrai...
Operating System & Kernel Study Guide-1 - converted.pdf
Model Code of Practice - Construction Work - 21102022 .pdf
IOT PPTs Week 10 Lecture Material.pptx of NPTEL Smart Cities contd
Well-logging-methods_new................

Algorithm explanations

  • 2. Kohonen’s self-organizing map (K-SOM) Artificial Neural Network - Used to reduce dimensionality and effectively represent n features of a given dataset into 2 features (n-d graph to 2-d graph)
  • 3. Kohonen’s Self-Organizing Map (SOM) • Each node’s weights are initialized • One vector (instance) from the dataset is taken • Every node is examined to calculate which one’s weights are most like the input vector. The winning node is commonly known as the Best Matching Unit (BMU), given by • Then the neighborhood of the BMU is calculated. The amount of neighbors decreases over time. • The winning weight is rewarded with becoming more like the sample vector. • The next vector is then taken from the dataset. These steps repeat till all data points are calibrated.
  • 5. K-Means Clustering (K-MC) Instance-based lazy learning - Used to cluster data such that similar data points are represented together and dissimilar data points are further away.
  • 6. K-means clustering (k-mc) • For k number of clusters, choose k random data points as centroids of the clusters. • Now choose one data point after another (other than the centroid points) and calculate distance (preferably Euclidean), given by • Cluster the data point with the nearest centroid and recalculate the centroid taking mean of coordinates • Repeat for other data points
  • 8. Logistic Regression (LR) Instance based quick learning - Used to map dependent variables to an independent variable and separate the positive and negative instances.
  • 9. Logistic regression (LR) • Take data point • Find class probability using sigmoid function as follows: • Then give the data a threshold such that if the value of g(z) is above that threshold then it is considered positive (1) otherwise it is considered negative (0)
  • 11. Support Vector machine (SVM) Instance-based quick learning - Used to make a hyperplane that separates the positive and negative samples with highest amount of margin
  • 12. Support Vector Machine (SVM) • The data points are first plotted in an n-dimensional space. • Many different lines, planes or hyperplanes (based on number of dimensions) are plotted to separate the two cases • The line/plane/hyperplane is chosen which has a maximum amount of margin between the positive and negative data points. This can be represented as • The above formula can be used to identify positive classes
  • 14. C4.5 decision tree (DT) Decision Tree based learning - Used for binary classification (two outcomes) in a question-answer tree format, with most relevant questions at the top and least relevant questions at the bottom
  • 15. C4.5 Decision tree (DT) • Find the feature of data that lends most information towards the outcome by using a method (generally least-error, information-gain, or gini coefficient) • Place that feature as the root node, and draw out branches, one for each of the values of that feature, and then assign child node as follows: • If that value of the node is decisive (i.e., if that value gives a decisive outcome) put the outcome as the leaf node • Else if the value is indecisive, repeat the above steps to calculate the sub-tree with the data used being the sub-dataset where that value exists
  • 17. Random forest (RF) Decision Tree based learning - An ensemble model that creates a lot of random trees and takes majority vote of their decisions
  • 18. Random forest (RF) • Create n decision trees randomly based on the features in the dataset, such that one tree may be a subset of another but no tree is the exact same • Obtain the result of all the decision trees and take majority vote to get the result
  • 20. Gradient boosting decision tree (GBdt) Decision Tree based greedy learning - An advanced decision tree that assigns weights to questions and calibrates weight to get a maximum accuracy tree.
  • 21. Gradient boosting decision tree (gbdt) • Build a decision tree similarly to c4.5 decision trees, but by assigning random initial weights to the given questions and sorting according to descending order of weights. • Assign a learning rate which defines how quick the tree will change. • Predict values for a data instance as (Prediction = Average value + learning rate x weighted increment) • Find the difference between the results (Difference = correct result - predicted value) • If the result is wrong, then adjust the weights for the next tree such that the overall result is closer to the result of that instance. • Take the summative results of all the given decision trees and classify accordingly
  • 23. K-nearest neighbors (knn) Instance based lazy learning - Simple, lazy learning algorithm which classifies the given data points according to the classes the nearest points.
  • 24. K-nearest neighbors (knn) • Plot all the known classified data points into an n-dimensional space • Consider a point whose dimensional coordinates are known (test point) but whose class is unknown • Compute distance as • Find the nearest 1 neighbor and consider its class • Continue considering more and more points such that the decision considers a lot of positive and negative class points with as much confidence as possible. • Use the most confident class as the class of the test point