SlideShare a Scribd company logo
ACEEE Int. J. on Network Security , Vol. 02, No. 04, Oct 2011



        Performance Evaluation of Classifiers used for
            Identification of Encryption Algorithms
                                            Suhaila Omer Sharif1, Saad P. Mansoor2
                                1,2
                                      School of Computer Science, Bangor University, Bangor, UK
                                                     1
                                                       s.o.sharif@bangor.ac.uk
                                                    2
                                                      s.mansoor@bangor.ac.uk


Abstract—Evaluating classifier performance is a critical                for Identification of a block cipher. The identification of
problem in pattern recognition and machine learning. In this            encryption method was considered as a document
paper pattern recognition techniques were applied to identify           categorization task. A.Cufoglu et al. [2] used pattern
encryption algorithms. Four different block cipher algorithms           recognition classifiers to identify user profiling. The results
were considered, DES, IDEA, AES, and RC2 operating in
                                                                        observed that NB classifier generate the best performance
(Electronic Codebook) ECB mode. Eight different classification
techniques were used for this purpose, these are: Naïve                 over user related information. In [3] a study was carried out
Bayesian (NB), Support Vector Machine (SVM), neural                     to use classifiers to identify the cipher encryption method.
network (MLP), Instance based learning (IBL), Bagging (Ba),             The simulation results show that, the RoFo classifier has the
AdaBoostM1 (MdaBM1), Rotation Forest (RoFo), and Decision               highest classification accuracy attributes by using tree
Tree (C4. 5). The result shows that using pattern recognition           structures obligatory on Naïve Bayesian. In this study a
is a useful technique to identify the encryption algorithm,             simulation test was carried to evaluate different types of
and according to our simulation using one encryption of key             classifiers in terms of classification accuracy of the encryption
provide better classification than using different keys.                data.
Furthermore, increase the number of the input files will
improve the accuracy.
                                                                                           II.   BASICS OF CLASSIFIERS

Index Terms—Pattern Recognition, Classifiers, Encryption,                  This section briefly explains the classifiers used in this
Cryptanalysis                                                           study.
                       I. INTRODUCTION                                  A. Naive Bayesian (NB)
    A typical cipher takes a plaintext message and some secret               Naïve Bayesian is one of the most computationally
keying data as its input and produces an encrypted version              straightforward and efficient Bayesian Classifier methods. It
of the original message known as the ciphertext. An attack              has been used as an effective classifier for many years. It has
on a cipher can make use of the ciphertext alone or it can              two advantages over many other classifiers. The first being,
make use of some plaintext and its corresponding ciphertext.            it is easy to construct, as the structure is given a priori thus,
Cryptanalysis is the process of recovering the plaintext and/           no structure learning procedure is required. Secondly, the
or key from a ciphertext. Most encryption algorithms have a             classification process is very efficient. This is due to the
finite key space and, hence, are vulnerable to an exhaustive            assumption that all the features are independent of each other.
key search attack. However, it is very difficult to identify the        It has surprisingly out performed many sophisticated
encryption keys because the size of the key is such that the            classifiers, especially where the features are not strongly
time and resources required are not generally available. A              correlated [4].
random search through a finite but large key space is not               B. Support Vector Machines (SVMs)
usually an acceptable cryptanalysts tool. In cryptanalysis                  This is a useful technique for data classification. It is
when only the ciphertext is obtained, there are initially two           based on statistical learning theory presented by V.N.Vapnik,
significant tasks, identification of the encryption technique           and has been productively applied to many classification
applied, and the encryption key identification. Statistical             and pattern recognition problems such as text classification
techniques and machine learning based techniques have been              and image recognition. These methods do classification by
applied to identify the encryption method from the encrypted            building an N-dimensional hyper plane that optimally divides
files. In this work, however, we are exploring the possibility          the data into two groups: the training and test data points.
of identifying encryption algorithm by applying pattern                 The objective of SVM method is to produce a model (based
recognition techniques. The statistical methods use the                 on the training data) which predicts the target values of the
frequency at which letters of the alphabet occur in the                 test data knowing only the test data attributes [5].
encrypted file. While in machine learning based methods,
the task of identification of the encryption method is                  C. Neural Networks (MLP)
considered as a pattern classification task. There are a number             These classifiers are widely used in many fields. They
of research papers that evaluate some of the classification             model the concept of neurons in neuroscience and include
algorithm. Dileep [1] suggests using a Support Vector Machine           sets of connected nodes along with the weights for nodes
© 2011 ACEEE                                                       42
DOI: 02.ICT.2011.02.60
ACEEE Int. J. on Network Security , Vol. 02, No. 04, Oct 2011


and arcs. Neural nets are self-organizing, adaptive and              for a base classifier, the feature set is randomly split into K
nonlinear. Neural network are versatile and powerful classifiers,    subsets (K is a parameter of the algorithm) and Principal
but they rely on a number of parameter choices to specify the        Component Analysis (PCA) is applied to each subset. All
network architecture and to control the training process [5,6].      principal components are retained in order to preserve the
                                                                     variability information in the data. Thus, K axis rotations take
D. Instance Based Learner (IBL)
                                                                     place to form the new features for a base classifier. The idea
    IBL generates classification predictions using only              behind the rotation approach is to encourage simultaneously
specific instances. Unlike Nearest Neighbour algorithm, IBL          individual accuracy and diversity within the ensemble.
normalizes its attributes’ ranges, processes instances               Diversity promoted through the feature extraction for each
incrementally, and has a simple policy for tolerating missing        base classifier. Decision trees were chosen here because they
values. IBL applies a simple Euclidean distance function to          are sensitive to rotation of the feature axes, hence the name
supply graded matches between training instances and given           “forest.” Accuracy is sought by keeping all principal
test instance [7]. Equation (1) represents the similarity that is    components and using the whole data set to train each base
used within the IBL algorithm.                                       classifier. A very useful characteristic of the proposed method
  Similarity(x, y) =                                        (1)      is that RoFo can be used in conjunction with practically any
                                                                     base classifies that are used in the ensemble creation process.
  Where n: is the number of instance attributes                      The thought of the rotation approach is to encourage
                                                                     simultaneously individual accuracy and diversity within the
The numeric value attributes are represented by (2).                 ensemble.

              f (xi, yi) = (xi – yi)2                   (2)          H. C4.5 Classification Algorithm
The Boolean and symbolic attributes are represented by (3).              This algorithm constructs a decision tree (DT) starting
                                                                     from a training set, which is a set of cases. Each case specifies
              f (xi, yi) = (xi   yi)                         (3)     values for a collection of attributes and for a class. The
                                                                     decision tree is used to classify (assign) a class value to a
E. Bagging (Ba)                                                      case depending on the values of the attributes of the case.
    This ensemble approach uses a predefined number of               The class specified at the leaf is the class predicted by the
classifier, each one trained on a bootstrap sample of training       decision tree. A performance measure of a decision tree over
data. Each classifier is trained on a set of n training examples,    a set of cases is the classification error which is the percentage
drawn randomly with replacement from the original training           of the misclassification cases whose predicted class differs
set of size n. Such a training set is called a bootstrap replicate   from actual class [12].
of the original set. Each bootstrap replicate contains, on
average, 63.2% of the original training set, with many examples                         III. EXPERMINT SETUP
appearing multiple times. [8, 9].
                                                                         Numbers of experiments were carried out to establish the
F. AdaBoost Algorithm Boosting                                       classifiers accuracy to identifying the following encryption
    The most popular classifier is AdaBoost, a powerful              algorithms AES, RC2, DES and IDEA, operating in electronic
classifier built with linear combination of member classifiers.      codebook (ECB) mode. In the simulation, different instances
The idea behind it is obtaining a highly accurate classifier by      were used (120, 240, and 400); these represent the total number
combining many weak classifiers. Each weak classifier is             of input files (ciphertext) used. The plaintext was encrypted
required to be moderately accurate, i.e. better than random          using 1, 3, 5, and 30 keys for each algorithm. Matlab® was
guessing. The classifiers in the ensemble are added one at a         used to extract an 8-bits histogram of the encrypted data;
time so that each subsequent classifier is trained on data,          this was then submitted to WEKA (Weikato Environment for
which have been “hard” for the previous ensemble members.            Knowledge Analysis) machine learning platform that supplies
A set of weights is maintained across the objects in the data        a work bench consisting of a collection of learning schemes,
set so that objects that have been difficult to classify acquire     which are applied for data mining and machine learning. The
more weight, forcing subsequent classifiers to focus on them.        primary learning schemes in Weka are “classifiers” and they
The ensemble construction through AdaBoost comes from                tempt a rule set or decision tree that models the data, in
theory and is, in fact, equivalent to a logistic regression model    addition Weka has different algorithms for learning
by a fitting an additive stage-wise estimation procedure. One        association rules clustering data, visualization, regression,
of the reasons why AdaBoost algorithm yields good results            and data processing.
is the diversity among the weak classifiers. Still, there is no
specific benchmark to determine diversity in the process [10,                         IV. SIMULATION RESULTS
11].                                                                 In this section the results of eight classifiers are com-
G. RotationForest (RoFo)                                         pared, based on the accuracy of the identification. The simu-
                                                                 lations conducted using four different block cipher algorithms
   This classifier is based on feature extraction and was
                                                                 with different key sizes. Each dataset have the same number
proposed by Rodriguez et al [10]. To create the training data
© 2011 ACEEE                                                  43
DOI: 02.ICT.2011.02.60
ACEEE Int. J. on Network Security , Vol. 02, No. 04, Oct 2011


of attributes and different number of instances, which varies             64-bits, IDEA 128-bits, RC2 128-bits we used 30, 60 and 100
from 120 to 400 instances. A 10 fold cross-validations was                input files, thus the total numbers of the instances are 120,
used as a test mode where 10 pair of training sets and testing            240 and 400 respectively. Fig 2 shows that for all classifiers
sets is created. All the classifiers run on the same training sets        using 400 input files achieve high accuracy and using 100
and have been tested on the same testing sets to obtain the               input files achieved lowest accuracy. It is also evident that
classification accuracy. To estimate a value on training data             RoFo classifier produces the best results for all instances,
set, we define an accuracy measure as shown in equation (4).               and IBL perform very badly when operating with 100 and
                                                                          240 instances. In final simulation the effect of the number of
      Accuracy (%) =                                          (4)
                                                                          keys used to encrypt the plaintext was investigated for the
In the first simulation, we have been using 240 input files               following encryption algorithm AES, RC2, and IDEA with fixed
with AES (128,192, and 256-bits), DES 64-bit, IDEA 128-bit,               128-bit encryption key and DES with 64-bits. The number of
and RC2 (42, 84, and128-bits) where each algorithm has 30                 input files used was 30 and the number of keys was 1, 3, 5 and
input files. The simulation was conducted using different                 30. Here, Fig 3 shows that RoFo classifier has better overall
numbers of keys for each algorithm; Fig 1 shows the calculated            performance and AdaBM1 achieved lowest accuracy. Fur-
accuracy for different set of data. It can be seen that, using 1          thermore, and as expected all the classifiers (apart from
encryption key produces high classification accuracy (100%)               AdaBM1) produce good accuracy when using 1, 3 and 5
meaning that the 240 instances (files) were correctly classified.         encryption keys. It can be seen as expected that the accu-
In contrast using 30 different encryptions keys (one for each             racy drops to 25 –55 when using one key per each file (30
file) result in lower classification accuracy. This is expected           keys).
as the keys are generated randomly and will affect the pattern
of test data. The results show that the accuracy of the
classification will be reduced with increased the number of
the encryption keys. The second simulation, deals with the
effect of increasing the number of input files (instances) on
overall accuracy. Here for each algorithm, AES 128-bits, DES




                                                                                Figure 3. Encryption accuracy for different key seizes




       Figure 1. Classification accuracy for each algorithm




                                                                                                        (a) All data




                                                                                                   (b) Class centers
       Figure 2. Classification accuracy for each algorithm                            Figure 4. Scatter plot for 240 data points

© 2011 ACEEE                                                         44
DOI: 02.ICT.2011.02. 60
ACEEE Int. J. on Network Security , Vol. 02, No. 04, Oct 2011


Fig 4, shows data scatter-plots and the centres of the 240                                          REFERENCES
data points. The plot shows that the classes are founded                 [1] A. Dileep and C. Chandra Sekhar, “Indemnification of Block
sporadically. The centres of the “clouds” of points for 8 classes        Ciphers using Support Vector Machines, “International Joint
were plotted in Fig 4-b, according to this plot, all algorithms          Conference on Neural Networks” Vancouver, Canada, pp. 2696–
have similar representation. Note that the scales of all plots           2701, July 16-21,2006.
are different. The class centres are indistinguishable if plot-          [2] A. Cufoglu, M. Lohi, K. Madani, “A Comparative Study of
ted on the axes of a subplot. Finally, we can say this high-             Selected Classifiers with Classification Accuracy in User Profiling”,
lights the difficulty in recognising the type of code through            World Congress on Computer Science and Information Engineering.
simple pattern classification methods. Fig 5 shows an image              Pp 708-712, San Francisco, 2009.
                                                                         [3] S.O. Sharif, L.I. Kuncheva, and S.P. Mansoor, “Classifying
of the distance matrix D computed by using identification
                                                                         Encryption Algorithms Using Pattern Recognition Techniques,”
accuracy of each encryption algorithm. The block of 30 by 30             Neural Networks, 3rd International Conference on Advanced
distances is outlining with black lines. Blue color means high           Computer Theory and Engineering (ICACTE), 2010
similarity while yellow and red indicate low similarity. The             [4] A.Cufoglu, M. Lohi, and K. Madani, “Classification Accuracy
class labels are as follows: 1 AES(128), 2 AES(192), 3 AES(256),         Performance of Naïve Bayesian (NB), Bayesian Networks (BN),
4 DES(64), 5 IDEA(128), 6 RC2(128),7 RC2(42), and 8 RC2(84).             Lazy Learning of Bayesian Rules (LBR) and Instance-Based Learner
The encoding technique that stands out from the rest is AES.             (IB1) - comparative study,” 2008 International Conference on
The 3-by-3 block sub-matrix in the top left corner is largely            Computer Engineering & Systems, 2008, pp. 210-215.
blue, shown the similarity within the code. Interestingly, AES           [5] L. I. Kuncheva., J. J. Rodríguez, “Classifier ensembles for
                                                                         fMRI data analysis: an experiment,” Magnetic resonance imaging,
algorithm is distinguishable from the rest of the algorithm
                                                                         vol. 28, no. 4, pp. 583-93, May. 2010.
also noted that the three versions of AES (128, 192, and 256)            [6] Y. Maeda and M. Wakamura, “Simultaneous Perturbation
are not distinguishable within AES. The red vertical and hori-           Learning rule for Recurrent neural Networks and its FPGA
zontal lines demonstrate the unusually large distances com-              implementation.,” IEEE transactions on neural networks a
pared to the rest. Unlike ECB, there is no clear pattern to              publication of the IEEE Neural Networks Council, vol. 16, 2005,
suggest that any of the codes are distinguishable.                       pp. 1664-1672.
                                                                         [7]David W. Aha, Kibler D. and K. Albert M. (1991) " Instance-
                                                                         based learning algorithms" . Machine Learning journal, Vol 1, No 6,
                                                                         ISDN 1573- 0565, pp. 37-66. [Online] Available from: http://
                                                                         www.springerlink.com/content/kn127378pg361187/fulltext.pdf
                                                                         [8] Q. He, F. Zhuang, X. Zhao, and Z. Shi, “Enhanced Algorithm
                                                                         Performance for Classification Based on Hyper Surface using
                                                                         Bagging and Adaboost,” 2007 International Conference on Machine
                                                                         Learning and Cybernetics, 2007, pp. 3624-3629.
                                                                         [9] S. Ruggieri, “Efficient C4.5” IEEE Transactions on Knowledge
                                                                         and Data Engineering, vol. 14, 2002, pp. 438-444.
                                                                         [10] J.J. Rodríguez, L.I. Kuncheva, and C.J. Alonso, “Rotation
                                                                         forest: A new Classifier Ensemble method,” IEEE transactions on
                                                                         pattern analysis and machine intelligence, vol. 28, 2006, pp. 1619-
                                                                         1630.
                                                                         [11] T.-K. An and M.-H. Kim, “A New Diverse AdaBoost
                                                                         Classifier,” 2010 International Conference on Artificial Intelligence
                   Figure 5. Distance matrix D
                                                                         and Computational Intelligence, pp. 359-363, Oct. 2010.
                                                                         [12] G. Stiglic and P. Kokol, “Effectiveness of Rotation Forest in
                                                                         Meta- learning Based Gene Expression Classification,” Symposium
                          CONCLUSIONS                                    a Quarterly Journal in Modern Foreign Literatures, 2007.
    In this paper, a simulation experiment was conducted to
identify encryption algorithm of encrypted data using variety
of classifies. The results show that pattern recognition
techniques, are useful tools for cryptanalysis as a means of
identifying the type of encryption algorithm used to encrypt
the data. This work shows that increasing the number of
encryption keys will result in reducing the classification
accuracy. Nevertheless, the result shows that it is possible
to achieve accuracy above 40% with some classifiers when
each file was encrypted with different key. It was also clear
that increase the number of files used will improve accuracy.




© 2011 ACEEE                                                        45
DOI: 02.ICT.2011.02. 60

More Related Content

PDF
Recognition of handwritten digits using rbf neural network
PDF
DataEngConf: Feature Extraction: Modern Questions and Challenges at Google
PDF
H42054550
PPTX
Representation and recognition of handwirten digits using deformable templates
PPTX
HANDWRITTEN DIGIT RECOGNITION USING k-NN CLASSIFIER
PDF
Neural Networks in the Wild: Handwriting Recognition
PDF
Segmentation and recognition of handwritten digit numeral string using a mult...
PPT
Intro to Deep learning - Autoencoders
Recognition of handwritten digits using rbf neural network
DataEngConf: Feature Extraction: Modern Questions and Challenges at Google
H42054550
Representation and recognition of handwirten digits using deformable templates
HANDWRITTEN DIGIT RECOGNITION USING k-NN CLASSIFIER
Neural Networks in the Wild: Handwriting Recognition
Segmentation and recognition of handwritten digit numeral string using a mult...
Intro to Deep learning - Autoencoders

What's hot (19)

PDF
Cv4201644655
PDF
Et25897899
PPTX
Autoencoders for image_classification
PDF
Steganography using Coefficient Replacement and Adaptive Scaling based on DTCWT
PDF
HIS'2008: Genetic-based Synthetic Data Sets for the Analysis of Classifiers B...
PDF
Artificial neural networks and its application
PDF
L026070074
PDF
Comparative Performance of Image Scrambling in Transform Domain using Sinusoi...
PPTX
Piotr Mirowski - Review Autoencoders (Deep Learning) - CIUUK14
PDF
Deep Style: Using Variational Auto-encoders for Image Generation
PDF
GENERAL REGRESSION NEURAL NETWORK BASED POS TAGGING FOR NEPALI TEXT
PDF
Practical deepllearningv1
PDF
SOIAM (SOINN-AM)
DOCX
Mcs 024 assignment solution (2020-21)
PDF
Dynamic thresholding on speech segmentation
PDF
A Secure & Optimized Data Hiding Technique Using DWT With PSNR Value
PPTX
Survey on Deep Neural Network Watermarking techniques
PDF
Study of Different Multi-instance Learning kNN Algorithms
PDF
Image Captioning Generator using Deep Machine Learning
Cv4201644655
Et25897899
Autoencoders for image_classification
Steganography using Coefficient Replacement and Adaptive Scaling based on DTCWT
HIS'2008: Genetic-based Synthetic Data Sets for the Analysis of Classifiers B...
Artificial neural networks and its application
L026070074
Comparative Performance of Image Scrambling in Transform Domain using Sinusoi...
Piotr Mirowski - Review Autoencoders (Deep Learning) - CIUUK14
Deep Style: Using Variational Auto-encoders for Image Generation
GENERAL REGRESSION NEURAL NETWORK BASED POS TAGGING FOR NEPALI TEXT
Practical deepllearningv1
SOIAM (SOINN-AM)
Mcs 024 assignment solution (2020-21)
Dynamic thresholding on speech segmentation
A Secure & Optimized Data Hiding Technique Using DWT With PSNR Value
Survey on Deep Neural Network Watermarking techniques
Study of Different Multi-instance Learning kNN Algorithms
Image Captioning Generator using Deep Machine Learning
Ad

Viewers also liked (20)

DOCX
Security evaluation of pattern classifiers under attack
PDF
Design of robust classifiers for adversarial environments - Systems, Man, and...
PDF
Battista Biggio, Invited Keynote @ AISec 2014 - On Learning and Recognition o...
PDF
Understanding the risk factors of learning in adversarial environments
PDF
Assessment of Market Power in Deregulated Electricity Market with Congestion
PDF
Energy Efficient Data Transmission through Relay Nodes in Wireless Sensor Net...
PDF
Neural Network Based Noise Identification in Digital Images
PDF
An Information Maximization approach of ICA for Gender Classification
PDF
FPGA Based Design of High Performance Decimator using DALUT Algorithm
PDF
Depth-Image-based Facial Analysis between Age Groups and Recognition of 3D Faces
PDF
Inverse Gamma Distribution based Delay and Slew Modeling for On- Chip VLSI RC...
PDF
An Iterative Improved k-means Clustering
PDF
User Controlled Privacy in Participatory Sensing
PDF
GR-FB Block Cleaning Scheme in Flash Memory
PDF
An Adaptive Energy Efficient Reliable Routing Protocol for Wireless Sensor Ne...
PDF
LabVIEW - Teaching Aid for Process Control
PDF
Black Box for Machine Tools; Based on Open CNC Architecture Control Systems
PDF
Genetic Algorithm based Optimal Placement of Distributed Generation Reducing ...
PDF
Interleaving Technique in Multiphase Buck & Boost Converter
PDF
A 2-tier Data Hiding Technique Using Exploiting Modification Direction Method...
Security evaluation of pattern classifiers under attack
Design of robust classifiers for adversarial environments - Systems, Man, and...
Battista Biggio, Invited Keynote @ AISec 2014 - On Learning and Recognition o...
Understanding the risk factors of learning in adversarial environments
Assessment of Market Power in Deregulated Electricity Market with Congestion
Energy Efficient Data Transmission through Relay Nodes in Wireless Sensor Net...
Neural Network Based Noise Identification in Digital Images
An Information Maximization approach of ICA for Gender Classification
FPGA Based Design of High Performance Decimator using DALUT Algorithm
Depth-Image-based Facial Analysis between Age Groups and Recognition of 3D Faces
Inverse Gamma Distribution based Delay and Slew Modeling for On- Chip VLSI RC...
An Iterative Improved k-means Clustering
User Controlled Privacy in Participatory Sensing
GR-FB Block Cleaning Scheme in Flash Memory
An Adaptive Energy Efficient Reliable Routing Protocol for Wireless Sensor Ne...
LabVIEW - Teaching Aid for Process Control
Black Box for Machine Tools; Based on Open CNC Architecture Control Systems
Genetic Algorithm based Optimal Placement of Distributed Generation Reducing ...
Interleaving Technique in Multiphase Buck & Boost Converter
A 2-tier Data Hiding Technique Using Exploiting Modification Direction Method...
Ad

Similar to Performance Evaluation of Classifiers used for Identification of Encryption Algorithms (20)

PDF
IJCSI-10-6-1-288-292
PDF
A Few Useful Things to Know about Machine Learning
PDF
Analysis of machine learning algorithms for character recognition: a case stu...
PPT
Textmining Predictive Models
PPT
Textmining Predictive Models
PPT
Textmining Predictive Models
PDF
Hypothesis on Different Data Mining Algorithms
DOC
Cyb 5675 class project final
PPTX
Statistical classification: A review on some techniques
PDF
Decision tree lecture 3
PDF
Supervised Machine Learning: A Review of Classification ...
PDF
Detection of malicious attacks by Meta classification algorithms
PDF
CLASSIFICATION ALGORITHM USING RANDOM CONCEPT ON A VERY LARGE DATA SET: A SURVEY
PDF
X24164167
PDF
Classifier Model using Artificial Neural Network
PDF
Classification Techniques: A Review
PPTX
UNIT 3: Data Warehousing and Data Mining
PDF
An Introduction to Supervised Machine Learning and Pattern Classification: Th...
PDF
IRJET- Performance Evaluation of Various Classification Algorithms
PDF
IRJET- Performance Evaluation of Various Classification Algorithms
IJCSI-10-6-1-288-292
A Few Useful Things to Know about Machine Learning
Analysis of machine learning algorithms for character recognition: a case stu...
Textmining Predictive Models
Textmining Predictive Models
Textmining Predictive Models
Hypothesis on Different Data Mining Algorithms
Cyb 5675 class project final
Statistical classification: A review on some techniques
Decision tree lecture 3
Supervised Machine Learning: A Review of Classification ...
Detection of malicious attacks by Meta classification algorithms
CLASSIFICATION ALGORITHM USING RANDOM CONCEPT ON A VERY LARGE DATA SET: A SURVEY
X24164167
Classifier Model using Artificial Neural Network
Classification Techniques: A Review
UNIT 3: Data Warehousing and Data Mining
An Introduction to Supervised Machine Learning and Pattern Classification: Th...
IRJET- Performance Evaluation of Various Classification Algorithms
IRJET- Performance Evaluation of Various Classification Algorithms

More from IDES Editor (20)

PDF
Power System State Estimation - A Review
PDF
Artificial Intelligence Technique based Reactive Power Planning Incorporating...
PDF
Design and Performance Analysis of Genetic based PID-PSS with SVC in a Multi-...
PDF
Optimal Placement of DG for Loss Reduction and Voltage Sag Mitigation in Radi...
PDF
Line Losses in the 14-Bus Power System Network using UPFC
PDF
Study of Structural Behaviour of Gravity Dam with Various Features of Gallery...
PDF
Assessing Uncertainty of Pushover Analysis to Geometric Modeling
PDF
Secure Multi-Party Negotiation: An Analysis for Electronic Payments in Mobile...
PDF
Selfish Node Isolation & Incentivation using Progressive Thresholds
PDF
Various OSI Layer Attacks and Countermeasure to Enhance the Performance of WS...
PDF
Responsive Parameter based an AntiWorm Approach to Prevent Wormhole Attack in...
PDF
Cloud Security and Data Integrity with Client Accountability Framework
PDF
Genetic Algorithm based Layered Detection and Defense of HTTP Botnet
PDF
Enhancing Data Storage Security in Cloud Computing Through Steganography
PDF
Low Energy Routing for WSN’s
PDF
Permutation of Pixels within the Shares of Visual Cryptography using KBRP for...
PDF
Rotman Lens Performance Analysis
PDF
Band Clustering for the Lossless Compression of AVIRIS Hyperspectral Images
PDF
Microelectronic Circuit Analogous to Hydrogen Bonding Network in Active Site ...
PDF
Texture Unit based Monocular Real-world Scene Classification using SOM and KN...
Power System State Estimation - A Review
Artificial Intelligence Technique based Reactive Power Planning Incorporating...
Design and Performance Analysis of Genetic based PID-PSS with SVC in a Multi-...
Optimal Placement of DG for Loss Reduction and Voltage Sag Mitigation in Radi...
Line Losses in the 14-Bus Power System Network using UPFC
Study of Structural Behaviour of Gravity Dam with Various Features of Gallery...
Assessing Uncertainty of Pushover Analysis to Geometric Modeling
Secure Multi-Party Negotiation: An Analysis for Electronic Payments in Mobile...
Selfish Node Isolation & Incentivation using Progressive Thresholds
Various OSI Layer Attacks and Countermeasure to Enhance the Performance of WS...
Responsive Parameter based an AntiWorm Approach to Prevent Wormhole Attack in...
Cloud Security and Data Integrity with Client Accountability Framework
Genetic Algorithm based Layered Detection and Defense of HTTP Botnet
Enhancing Data Storage Security in Cloud Computing Through Steganography
Low Energy Routing for WSN’s
Permutation of Pixels within the Shares of Visual Cryptography using KBRP for...
Rotman Lens Performance Analysis
Band Clustering for the Lossless Compression of AVIRIS Hyperspectral Images
Microelectronic Circuit Analogous to Hydrogen Bonding Network in Active Site ...
Texture Unit based Monocular Real-world Scene Classification using SOM and KN...

Recently uploaded (20)

PDF
Agricultural_Statistics_at_a_Glance_2022_0.pdf
PDF
How UI/UX Design Impacts User Retention in Mobile Apps.pdf
PPTX
A Presentation on Artificial Intelligence
PDF
The Rise and Fall of 3GPP – Time for a Sabbatical?
PDF
CIFDAQ's Market Insight: SEC Turns Pro Crypto
PPTX
Effective Security Operations Center (SOC) A Modern, Strategic, and Threat-In...
PPTX
Detection-First SIEM: Rule Types, Dashboards, and Threat-Informed Strategy
PDF
NewMind AI Monthly Chronicles - July 2025
PPTX
Understanding_Digital_Forensics_Presentation.pptx
PPTX
Digital-Transformation-Roadmap-for-Companies.pptx
PDF
7 ChatGPT Prompts to Help You Define Your Ideal Customer Profile.pdf
DOCX
The AUB Centre for AI in Media Proposal.docx
PPT
Teaching material agriculture food technology
PDF
Dropbox Q2 2025 Financial Results & Investor Presentation
PDF
Bridging biosciences and deep learning for revolutionary discoveries: a compr...
PDF
Spectral efficient network and resource selection model in 5G networks
PDF
Shreyas Phanse Resume: Experienced Backend Engineer | Java • Spring Boot • Ka...
PPTX
PA Analog/Digital System: The Backbone of Modern Surveillance and Communication
PDF
Electronic commerce courselecture one. Pdf
PDF
Encapsulation theory and applications.pdf
Agricultural_Statistics_at_a_Glance_2022_0.pdf
How UI/UX Design Impacts User Retention in Mobile Apps.pdf
A Presentation on Artificial Intelligence
The Rise and Fall of 3GPP – Time for a Sabbatical?
CIFDAQ's Market Insight: SEC Turns Pro Crypto
Effective Security Operations Center (SOC) A Modern, Strategic, and Threat-In...
Detection-First SIEM: Rule Types, Dashboards, and Threat-Informed Strategy
NewMind AI Monthly Chronicles - July 2025
Understanding_Digital_Forensics_Presentation.pptx
Digital-Transformation-Roadmap-for-Companies.pptx
7 ChatGPT Prompts to Help You Define Your Ideal Customer Profile.pdf
The AUB Centre for AI in Media Proposal.docx
Teaching material agriculture food technology
Dropbox Q2 2025 Financial Results & Investor Presentation
Bridging biosciences and deep learning for revolutionary discoveries: a compr...
Spectral efficient network and resource selection model in 5G networks
Shreyas Phanse Resume: Experienced Backend Engineer | Java • Spring Boot • Ka...
PA Analog/Digital System: The Backbone of Modern Surveillance and Communication
Electronic commerce courselecture one. Pdf
Encapsulation theory and applications.pdf

Performance Evaluation of Classifiers used for Identification of Encryption Algorithms

  • 1. ACEEE Int. J. on Network Security , Vol. 02, No. 04, Oct 2011 Performance Evaluation of Classifiers used for Identification of Encryption Algorithms Suhaila Omer Sharif1, Saad P. Mansoor2 1,2 School of Computer Science, Bangor University, Bangor, UK 1 s.o.sharif@bangor.ac.uk 2 s.mansoor@bangor.ac.uk Abstract—Evaluating classifier performance is a critical for Identification of a block cipher. The identification of problem in pattern recognition and machine learning. In this encryption method was considered as a document paper pattern recognition techniques were applied to identify categorization task. A.Cufoglu et al. [2] used pattern encryption algorithms. Four different block cipher algorithms recognition classifiers to identify user profiling. The results were considered, DES, IDEA, AES, and RC2 operating in observed that NB classifier generate the best performance (Electronic Codebook) ECB mode. Eight different classification techniques were used for this purpose, these are: Naïve over user related information. In [3] a study was carried out Bayesian (NB), Support Vector Machine (SVM), neural to use classifiers to identify the cipher encryption method. network (MLP), Instance based learning (IBL), Bagging (Ba), The simulation results show that, the RoFo classifier has the AdaBoostM1 (MdaBM1), Rotation Forest (RoFo), and Decision highest classification accuracy attributes by using tree Tree (C4. 5). The result shows that using pattern recognition structures obligatory on Naïve Bayesian. In this study a is a useful technique to identify the encryption algorithm, simulation test was carried to evaluate different types of and according to our simulation using one encryption of key classifiers in terms of classification accuracy of the encryption provide better classification than using different keys. data. Furthermore, increase the number of the input files will improve the accuracy. II. BASICS OF CLASSIFIERS Index Terms—Pattern Recognition, Classifiers, Encryption, This section briefly explains the classifiers used in this Cryptanalysis study. I. INTRODUCTION A. Naive Bayesian (NB) A typical cipher takes a plaintext message and some secret Naïve Bayesian is one of the most computationally keying data as its input and produces an encrypted version straightforward and efficient Bayesian Classifier methods. It of the original message known as the ciphertext. An attack has been used as an effective classifier for many years. It has on a cipher can make use of the ciphertext alone or it can two advantages over many other classifiers. The first being, make use of some plaintext and its corresponding ciphertext. it is easy to construct, as the structure is given a priori thus, Cryptanalysis is the process of recovering the plaintext and/ no structure learning procedure is required. Secondly, the or key from a ciphertext. Most encryption algorithms have a classification process is very efficient. This is due to the finite key space and, hence, are vulnerable to an exhaustive assumption that all the features are independent of each other. key search attack. However, it is very difficult to identify the It has surprisingly out performed many sophisticated encryption keys because the size of the key is such that the classifiers, especially where the features are not strongly time and resources required are not generally available. A correlated [4]. random search through a finite but large key space is not B. Support Vector Machines (SVMs) usually an acceptable cryptanalysts tool. In cryptanalysis This is a useful technique for data classification. It is when only the ciphertext is obtained, there are initially two based on statistical learning theory presented by V.N.Vapnik, significant tasks, identification of the encryption technique and has been productively applied to many classification applied, and the encryption key identification. Statistical and pattern recognition problems such as text classification techniques and machine learning based techniques have been and image recognition. These methods do classification by applied to identify the encryption method from the encrypted building an N-dimensional hyper plane that optimally divides files. In this work, however, we are exploring the possibility the data into two groups: the training and test data points. of identifying encryption algorithm by applying pattern The objective of SVM method is to produce a model (based recognition techniques. The statistical methods use the on the training data) which predicts the target values of the frequency at which letters of the alphabet occur in the test data knowing only the test data attributes [5]. encrypted file. While in machine learning based methods, the task of identification of the encryption method is C. Neural Networks (MLP) considered as a pattern classification task. There are a number These classifiers are widely used in many fields. They of research papers that evaluate some of the classification model the concept of neurons in neuroscience and include algorithm. Dileep [1] suggests using a Support Vector Machine sets of connected nodes along with the weights for nodes © 2011 ACEEE 42 DOI: 02.ICT.2011.02.60
  • 2. ACEEE Int. J. on Network Security , Vol. 02, No. 04, Oct 2011 and arcs. Neural nets are self-organizing, adaptive and for a base classifier, the feature set is randomly split into K nonlinear. Neural network are versatile and powerful classifiers, subsets (K is a parameter of the algorithm) and Principal but they rely on a number of parameter choices to specify the Component Analysis (PCA) is applied to each subset. All network architecture and to control the training process [5,6]. principal components are retained in order to preserve the variability information in the data. Thus, K axis rotations take D. Instance Based Learner (IBL) place to form the new features for a base classifier. The idea IBL generates classification predictions using only behind the rotation approach is to encourage simultaneously specific instances. Unlike Nearest Neighbour algorithm, IBL individual accuracy and diversity within the ensemble. normalizes its attributes’ ranges, processes instances Diversity promoted through the feature extraction for each incrementally, and has a simple policy for tolerating missing base classifier. Decision trees were chosen here because they values. IBL applies a simple Euclidean distance function to are sensitive to rotation of the feature axes, hence the name supply graded matches between training instances and given “forest.” Accuracy is sought by keeping all principal test instance [7]. Equation (1) represents the similarity that is components and using the whole data set to train each base used within the IBL algorithm. classifier. A very useful characteristic of the proposed method Similarity(x, y) = (1) is that RoFo can be used in conjunction with practically any base classifies that are used in the ensemble creation process. Where n: is the number of instance attributes The thought of the rotation approach is to encourage simultaneously individual accuracy and diversity within the The numeric value attributes are represented by (2). ensemble. f (xi, yi) = (xi – yi)2 (2) H. C4.5 Classification Algorithm The Boolean and symbolic attributes are represented by (3). This algorithm constructs a decision tree (DT) starting from a training set, which is a set of cases. Each case specifies f (xi, yi) = (xi yi) (3) values for a collection of attributes and for a class. The decision tree is used to classify (assign) a class value to a E. Bagging (Ba) case depending on the values of the attributes of the case. This ensemble approach uses a predefined number of The class specified at the leaf is the class predicted by the classifier, each one trained on a bootstrap sample of training decision tree. A performance measure of a decision tree over data. Each classifier is trained on a set of n training examples, a set of cases is the classification error which is the percentage drawn randomly with replacement from the original training of the misclassification cases whose predicted class differs set of size n. Such a training set is called a bootstrap replicate from actual class [12]. of the original set. Each bootstrap replicate contains, on average, 63.2% of the original training set, with many examples III. EXPERMINT SETUP appearing multiple times. [8, 9]. Numbers of experiments were carried out to establish the F. AdaBoost Algorithm Boosting classifiers accuracy to identifying the following encryption The most popular classifier is AdaBoost, a powerful algorithms AES, RC2, DES and IDEA, operating in electronic classifier built with linear combination of member classifiers. codebook (ECB) mode. In the simulation, different instances The idea behind it is obtaining a highly accurate classifier by were used (120, 240, and 400); these represent the total number combining many weak classifiers. Each weak classifier is of input files (ciphertext) used. The plaintext was encrypted required to be moderately accurate, i.e. better than random using 1, 3, 5, and 30 keys for each algorithm. Matlab® was guessing. The classifiers in the ensemble are added one at a used to extract an 8-bits histogram of the encrypted data; time so that each subsequent classifier is trained on data, this was then submitted to WEKA (Weikato Environment for which have been “hard” for the previous ensemble members. Knowledge Analysis) machine learning platform that supplies A set of weights is maintained across the objects in the data a work bench consisting of a collection of learning schemes, set so that objects that have been difficult to classify acquire which are applied for data mining and machine learning. The more weight, forcing subsequent classifiers to focus on them. primary learning schemes in Weka are “classifiers” and they The ensemble construction through AdaBoost comes from tempt a rule set or decision tree that models the data, in theory and is, in fact, equivalent to a logistic regression model addition Weka has different algorithms for learning by a fitting an additive stage-wise estimation procedure. One association rules clustering data, visualization, regression, of the reasons why AdaBoost algorithm yields good results and data processing. is the diversity among the weak classifiers. Still, there is no specific benchmark to determine diversity in the process [10, IV. SIMULATION RESULTS 11]. In this section the results of eight classifiers are com- G. RotationForest (RoFo) pared, based on the accuracy of the identification. The simu- lations conducted using four different block cipher algorithms This classifier is based on feature extraction and was with different key sizes. Each dataset have the same number proposed by Rodriguez et al [10]. To create the training data © 2011 ACEEE 43 DOI: 02.ICT.2011.02.60
  • 3. ACEEE Int. J. on Network Security , Vol. 02, No. 04, Oct 2011 of attributes and different number of instances, which varies 64-bits, IDEA 128-bits, RC2 128-bits we used 30, 60 and 100 from 120 to 400 instances. A 10 fold cross-validations was input files, thus the total numbers of the instances are 120, used as a test mode where 10 pair of training sets and testing 240 and 400 respectively. Fig 2 shows that for all classifiers sets is created. All the classifiers run on the same training sets using 400 input files achieve high accuracy and using 100 and have been tested on the same testing sets to obtain the input files achieved lowest accuracy. It is also evident that classification accuracy. To estimate a value on training data RoFo classifier produces the best results for all instances, set, we define an accuracy measure as shown in equation (4). and IBL perform very badly when operating with 100 and 240 instances. In final simulation the effect of the number of Accuracy (%) = (4) keys used to encrypt the plaintext was investigated for the In the first simulation, we have been using 240 input files following encryption algorithm AES, RC2, and IDEA with fixed with AES (128,192, and 256-bits), DES 64-bit, IDEA 128-bit, 128-bit encryption key and DES with 64-bits. The number of and RC2 (42, 84, and128-bits) where each algorithm has 30 input files used was 30 and the number of keys was 1, 3, 5 and input files. The simulation was conducted using different 30. Here, Fig 3 shows that RoFo classifier has better overall numbers of keys for each algorithm; Fig 1 shows the calculated performance and AdaBM1 achieved lowest accuracy. Fur- accuracy for different set of data. It can be seen that, using 1 thermore, and as expected all the classifiers (apart from encryption key produces high classification accuracy (100%) AdaBM1) produce good accuracy when using 1, 3 and 5 meaning that the 240 instances (files) were correctly classified. encryption keys. It can be seen as expected that the accu- In contrast using 30 different encryptions keys (one for each racy drops to 25 –55 when using one key per each file (30 file) result in lower classification accuracy. This is expected keys). as the keys are generated randomly and will affect the pattern of test data. The results show that the accuracy of the classification will be reduced with increased the number of the encryption keys. The second simulation, deals with the effect of increasing the number of input files (instances) on overall accuracy. Here for each algorithm, AES 128-bits, DES Figure 3. Encryption accuracy for different key seizes Figure 1. Classification accuracy for each algorithm (a) All data (b) Class centers Figure 2. Classification accuracy for each algorithm Figure 4. Scatter plot for 240 data points © 2011 ACEEE 44 DOI: 02.ICT.2011.02. 60
  • 4. ACEEE Int. J. on Network Security , Vol. 02, No. 04, Oct 2011 Fig 4, shows data scatter-plots and the centres of the 240 REFERENCES data points. The plot shows that the classes are founded [1] A. Dileep and C. Chandra Sekhar, “Indemnification of Block sporadically. The centres of the “clouds” of points for 8 classes Ciphers using Support Vector Machines, “International Joint were plotted in Fig 4-b, according to this plot, all algorithms Conference on Neural Networks” Vancouver, Canada, pp. 2696– have similar representation. Note that the scales of all plots 2701, July 16-21,2006. are different. The class centres are indistinguishable if plot- [2] A. Cufoglu, M. Lohi, K. Madani, “A Comparative Study of ted on the axes of a subplot. Finally, we can say this high- Selected Classifiers with Classification Accuracy in User Profiling”, lights the difficulty in recognising the type of code through World Congress on Computer Science and Information Engineering. simple pattern classification methods. Fig 5 shows an image Pp 708-712, San Francisco, 2009. [3] S.O. Sharif, L.I. Kuncheva, and S.P. Mansoor, “Classifying of the distance matrix D computed by using identification Encryption Algorithms Using Pattern Recognition Techniques,” accuracy of each encryption algorithm. The block of 30 by 30 Neural Networks, 3rd International Conference on Advanced distances is outlining with black lines. Blue color means high Computer Theory and Engineering (ICACTE), 2010 similarity while yellow and red indicate low similarity. The [4] A.Cufoglu, M. Lohi, and K. Madani, “Classification Accuracy class labels are as follows: 1 AES(128), 2 AES(192), 3 AES(256), Performance of Naïve Bayesian (NB), Bayesian Networks (BN), 4 DES(64), 5 IDEA(128), 6 RC2(128),7 RC2(42), and 8 RC2(84). Lazy Learning of Bayesian Rules (LBR) and Instance-Based Learner The encoding technique that stands out from the rest is AES. (IB1) - comparative study,” 2008 International Conference on The 3-by-3 block sub-matrix in the top left corner is largely Computer Engineering & Systems, 2008, pp. 210-215. blue, shown the similarity within the code. Interestingly, AES [5] L. I. Kuncheva., J. J. Rodríguez, “Classifier ensembles for fMRI data analysis: an experiment,” Magnetic resonance imaging, algorithm is distinguishable from the rest of the algorithm vol. 28, no. 4, pp. 583-93, May. 2010. also noted that the three versions of AES (128, 192, and 256) [6] Y. Maeda and M. Wakamura, “Simultaneous Perturbation are not distinguishable within AES. The red vertical and hori- Learning rule for Recurrent neural Networks and its FPGA zontal lines demonstrate the unusually large distances com- implementation.,” IEEE transactions on neural networks a pared to the rest. Unlike ECB, there is no clear pattern to publication of the IEEE Neural Networks Council, vol. 16, 2005, suggest that any of the codes are distinguishable. pp. 1664-1672. [7]David W. Aha, Kibler D. and K. Albert M. (1991) " Instance- based learning algorithms" . Machine Learning journal, Vol 1, No 6, ISDN 1573- 0565, pp. 37-66. [Online] Available from: http:// www.springerlink.com/content/kn127378pg361187/fulltext.pdf [8] Q. He, F. Zhuang, X. Zhao, and Z. Shi, “Enhanced Algorithm Performance for Classification Based on Hyper Surface using Bagging and Adaboost,” 2007 International Conference on Machine Learning and Cybernetics, 2007, pp. 3624-3629. [9] S. Ruggieri, “Efficient C4.5” IEEE Transactions on Knowledge and Data Engineering, vol. 14, 2002, pp. 438-444. [10] J.J. Rodríguez, L.I. Kuncheva, and C.J. Alonso, “Rotation forest: A new Classifier Ensemble method,” IEEE transactions on pattern analysis and machine intelligence, vol. 28, 2006, pp. 1619- 1630. [11] T.-K. An and M.-H. Kim, “A New Diverse AdaBoost Classifier,” 2010 International Conference on Artificial Intelligence Figure 5. Distance matrix D and Computational Intelligence, pp. 359-363, Oct. 2010. [12] G. Stiglic and P. Kokol, “Effectiveness of Rotation Forest in Meta- learning Based Gene Expression Classification,” Symposium CONCLUSIONS a Quarterly Journal in Modern Foreign Literatures, 2007. In this paper, a simulation experiment was conducted to identify encryption algorithm of encrypted data using variety of classifies. The results show that pattern recognition techniques, are useful tools for cryptanalysis as a means of identifying the type of encryption algorithm used to encrypt the data. This work shows that increasing the number of encryption keys will result in reducing the classification accuracy. Nevertheless, the result shows that it is possible to achieve accuracy above 40% with some classifiers when each file was encrypted with different key. It was also clear that increase the number of files used will improve accuracy. © 2011 ACEEE 45 DOI: 02.ICT.2011.02. 60