SlideShare a Scribd company logo
International Journal of Engineering Research and Development
e-ISSN: 2278-067X, p-ISSN: 2278-800X, www.ijerd.com
Volume 7, Issue 6 (June 2013), PP. 57-61
57
A Review Paper on Feature Selection Methodologies and
Their Applications
Ms. Shweta Srivastava1
, Ms. Nikita Joshi2
, Ms. Madhvi Gaur3
Assistant Professor(CSE Department)
ABES Engineering College, Ghaziabad, India.
Abstract:- Feature selection is the process of eliminating features from the data set that are irrelevant with
respect to the task to be performed. Feature selection is important for many reasons such as simplification,
performance, computational efficiency and feature interpretability. It can be applied to both supervised and
unsupervised learning methodologies. Such techniques are able in improving the efficiency of various machine
learning algorithms and that of training as well. Feature selection speed up the run time of learning, improves
data quality and data understanding.
Keywords:- Feature Selection, Supervised, Search strategies, Unsupervised.
I. INTRODUCTION
Feature selection is used to satisfy the common goal of maximizing the accuracy of the classifier,
minimizing the related measurement costs; improve accuracy by reducing irrelevant and possibly redundant
features; reduce the complexity and the associated computational cost; and improve the probability that a
solution will be comprehensible and realistic [13]. Feature selection is one of the stage for preprocessing the
data to reduce the dimensionality [10]. It selects a subset of the existing features without any transformation. It
can be said as a special case of feature extraction.
II. FEATURE SELECTION METHODOLOGIES
When we use feature selection, smaller number of features are extracted which means fewer model
parameters. It improves the generalization capabilities and reduces complexity and execution time. Feature
Selection methodologies can be categorized as: supervised and unsupervised feature selection methods.
Supervised Feature Selection Method: A feature selection method requires a search strategy to select the
subset of attributes and an objective function to evaluate the selected subset of features.
Objective function is divided in three categories:
1. Filters: It consists of algorithms that are built in the adaptive systems for data analysis (predictors) [9]. They
use an evaluation function that relies on properties of the data. Distance based and margin based criterion can be
used for filters.
Fig.1. Filter method
2. Wrapper: The algorithms of this method are wrapped around the adaptive systems providing them subsets of
features and receiving their feedback (usually accuracy). These wrapper approaches are aimed at improving
A Review Paper on Feature Selection Methodologies and Their Applications
58
results of the specific predictors they work with [9]. It utilizes the classifier as black box to find the subset of
features based on their predictive power [14].
3. Embedded: They are based on performance evaluation metric calculated directly from the data, without direct
reference to the results of any data analysis systems.
Fig.2. Wrapper method
3. Embedded: They are based on performance evaluation metric calculated directly from the data, without direct
reference to the results of any data analysis systems.
Filter Methods Wrapper methods Embedded methods
Filter methods appear to be a
probably less optimal.[4]
Wrapper methods are superior
alternative in supervised learning
problems. [4]
The performance of embedded
method degrades if more irrelevant
features are inserted in target set.
Filter methods execute faster than
wrapper methods.
Wrapper methods execute slower
than filter methods.
Embedded method is faster than
the wrapper methods.
The result of filter method exhibits
more generality than that of
wrapper approach.
There is lack of generality in
wrapper methods as it is tied to
some classifier.
It also lacks generality as it is also
dependent on some classification
algorithm.
Filter method has more tendencies
to select large subset of data.
Wrapper method is more accurate
than filter methods as it achieves
better recognition rates rather than
that of filters.
They are least prone to over fitting.
Computational cost is less for large
data set in filter method.
Computational cost is more for
large data set in filter method.
Computational cost is less in
comparison to wrapper methods.
Independent of classification
algorithm.
Dependent on classification
algorithm.
Dependent on classification
algorithm.
Table 1: Comparison of filter, wrapper and embedded methods
Search Strategy:
A search strategy is required to select the candidate subsets and objective function evaluates each of the
candidate subset. Search strategy involves exhaustive, heuristic and randomized searching algorithms [15][16].
The time complexity is exponential in terms of dimensionality for exhaustive search and quadratic for heuristic
search. The complexity can be linear to the number of iterations in a random search [15]. There are several
hybrid algorithms also.
Exhaustive search: It evaluates a number of subsets that grows exponentially with the dimensionality of the
search space. It finds the 2N
combinations of all m features. Each of the subset is evaluated by objective function
and a measure of goodness is returned to the search algorithm. Branch and bound based algorithms are used for
A Review Paper on Feature Selection Methodologies and Their Applications
59
exhaustive search. There are several versions of branch and bound algorithm i.e. Basic BB (slowest), Enhanced
BB (fastest), Fast BB, BB with partial prediction [17].
Heuristic search: The number of subsets evaluated by BFF is less (even much less) than that needed by the
branch and bound algorithm [18]. They require large computational time. Heuristic search is problematic when
the data has highly correlated features. Heuristics help to reduce the number of alternatives from an exponential
number to a polynomial number [19]. For example: Sequential forward selection, sequential backward
elimination and bidirectional search.
Forward Selection: Forward selection considers the subset to be empty initially and keeps on adding one
feature at a time until the best feature subset is obtained.
Backward Elimination: Backward selection takes complete set of features as input and keeps on removing one
attribute at a time until the most appropriate subset of features is obtained.
Randomized search strategy: It performs randomized exploration of the search space where next direction is a
sample from a given probability [20]. For example: genetic algorithm.
Accuracy Complexity Advantages Disadvantages
Exhaustive Search It always finds
the best possible
solution.
Exponential It is highly
accurate.
Complexity is
high.
Heuristic Search This technique
is good if no
backtracking is
needed.
Quadratic It is simple and
fast.
Backtracking is
not possible.
Randomized
Search
It is good with
proper control
parameters.
Generally low It is designed to
escape local
minima.
It is difficult to
choose good
parameters.
Table 2: Comparison of Exhaustive, Heuristic and Randomized Search
Unsupervised feature selection:
The procedure of arranging the objects into natural classes whose members are similar to each other,
identified by a given metrics. Unsupervised feature selection is particularly difficult due to the absence of class
labels for feature relevance estimation. Unsupervised feature selection is a less constrained search problem
without class labels, depending on clustering quality measures and can achieve many equally valid feature
subsets. Feature redundancy and relevance is measured with respect to clusters instead of classes. The objective
of clustering is to maximize intra cluster similarity and minimizing inter cluster similarity. For example:
Agglomerative and partitioned clustering. Filter and wrapper approach are helpful in unsupervised feature
selection also.
III. APPLICATION OF FEATURE SELECTION IN VARIOUS FIELDS
Text classification:
There are various challenges related to automated text classification such as:
1. An appropriate data structure is to be selected to represent the documents.
2. An appropriate objective function is to be chosen to optimize to avoid overfitting and obtain good
generalization. Along with it algorithmic issues arising as a result of the high formal dimensionality of the data
are to be dealt with [1].
Genre classification:
Metadata such as filename, author, size, date, track length and genres are the common features used to
classify and retrieve genre documents. On the basis of these data, the classification is infeasible, so the feature
selection step is required. In case of genre classification feature selection is a process where a segment of an
audio is characterized into a compact numerical representation [2]. Feature selection is done to reduce the
dimensionality of the data as a preprocessing step prior to classification due to high dimensionality of the feature
sets.
Microarray data analysis:
1. Almost all bioinformatics problems have the number of features significantly larger than the number of
samples (high feature to sample ratio datasets) [3]. For example: Breast cancer classification on the basis of
microarray data. Though the information about all the genes is not required in case of such classification.
2. Content analysis and signal analysis in genomics also require feature selection.
A Review Paper on Feature Selection Methodologies and Their Applications
60
Software defect prediction:
There are various software quality assurance attributes such as reliability, functionality, fault
proneness, reusability, comprehensibility etc [6]. It is a critical issue to select most appropriate software metrics
that likely to indicate fault proneness.
Sentiment analysis:
Sentiment analysis is capturing favorability using natural language processing. It is not just a topic
based categorization. It deals with the computational treatment of opinion, sentiment, and subjectivity in text. It
is useful in recommendation systems and question answering [8]. To decide about the positivity or negativity of
the opinion on the basis of various features such as term presence, feature frequency, feature presence, term
position, POS tags, syntax, topic and negation etc. All of the features are not required in each and every case. So
feature selection need to be performed.
Stock market analysis:
There are hundreds of stock index futures. Along with it financial data including the stock market data
is too complex to be searched easily [11]. In particular, the existence of large amount of continuous data may
cause a challenging task to explicit concepts extraction from the raw data due to the huge amount of data space
determined by continuous features [12]. So it is necessary to reduce the dimensionality of data and irrelevant
factors before searching.
Image Retrieval
Feature selection is applied to content based image retrieval to allow efficient browsing, searching and
retrieving [22]. Content based image retrieval is to index the images on the basis of their own visual contents
(i.e. color, shape, texture etc.) instead of text based keyword indexing. The biggest problem for content
based image retrieval is large amount of images in database [21].
IV. CONCLUSION
This paper provides a comprehensive overview of various characteristic of feature selection. Feature
selection as a preprocessing step in very large databases collected from various applications. More work is
required to overcome limitations imposed as it is costly to visit high dimensionality data multiple times or
accessing instances at random times. In unsupervised feature selection, several clusters may exist in different
subspaces of small-small dimensionality, with their sets of dimensions overlapped or non-overlapped. As most
existing feature selection algorithms have quadratic or higher time complexity about N, so it is quite difficult to
handle high dimensionality. Therefore, more efficient search strategies and evaluation criteria are needed for
feature selection with large dimensionality.
REFERENCES
[1]. Anirban Dasgupta,Petros Drineas, Boulos Harb, Vanja Josifovski, Michael W. Mahoney; “Feature
Selection Methods for Text Classification”,KDD- ACM.,2007.
[2]. Shyamala Doraisamy, Shahram Golzari, Noris Mohd. Norowi, Md. Nasir B Sulaiman, Nur Izura Udzir;
” A Study on Feature Selection and Classification Techniques for Automatic Genre Classification of
Traditional Malay Music” Proceedings of ISMIR, 2008.
[3]. Gianluca Bontempi, Benjamin Haibe-Kains; “Feature selection methods for mining bioinformatics
data”,
[4]. Luis Talavera; “An evaluation of filter and wrapper methods for feature selection in categorical
clustering”, Proceedings of the 6th international conference on Advances in Intelligent Data Analysis,
2005.
[5]. Yvan Saeys, Inaki Inza, Pedro Larranaga; ”A review of feature selection techniques in bioinformatics”,
Oxford Journals, 2007.
[6]. N.Gayatri, S.Nickolas, A.V.Reddy; “Feature Selection Using Decision Tree Induction in Class level
Metrics Dataset for Software Defect Predictions”; Proceedings of the World Congress on Engineering
and Computer Science ; Vol I; 2010.
[7]. Tim O’Keefe, Irena Koprinska;” Feature Selection andWeighting Methods in Sentiment Analysis”;
Proceedings of the 14th Australasian Document Computing Symposium, Sydney, Australia; 2009.
[8]. Bo Pang, Lillian Lee; ”Opinion Mining and Sentiment Analysis”; Foundations and Trends in
Information Retrieval, Vol. 2, Nos. 1–2 pp- 1–135, 2008.
[9]. Wlodzislaw Duch; “Filter methods”, Springer- Feature Extraction Studies in Fuzziness and Soft
Computing Volume 207, 2006, pp 89-117.
A Review Paper on Feature Selection Methodologies and Their Applications
61
[10]. Chih- Fong Tsai; “Data pre-processing by genetic algorithms for bankruptcy prediction”, IEEE
International Conference on Industrial Engineering and Engineering Management, Pages- 1780 - 1783 ,
2011.
[11]. Kyoung-jae Kim, Ingoo Han; “Genetic algorithms approach to feature discretization in artificial neural
networks for the prediction of stock price index”; Expert Systems with Applications, Elsevier Science
Ltd; 2000.
[12]. Liu, H., & Setiono, R.; “Dimensionality reduction via discretization”; Knowledge-Based Systems, 9
(1), 67–72; 1996.
[13]. Steppe. J., K.W. Bauer, “Feature Saliency Measures”, Computers & Mathematics with Applications,
Vol 33, No. 8, pp. 109-126; 1997.
[14]. L.Ladha et al., “Feature Selection Methods and Algorithms”; International Journal on Computer
Science and Engineering (IJCSE), Vol. 3. No. 5, pp. 1787-1797; 2011.
[15]. Roberto Ruiz, Jos´e C. Riquelme, and Jes´us S. Aguilar-Ruiz; “Heuristic Search over a Ranking for
Feature Selection”; IWANN, LNCS 3512, pp. 742–749; 2005.
[16]. Yao-Hong Chan; “Empirical comparison of forward and backward search strategies in L-GEM based
feature selection with RBFNN”; International Conference on Machine Learning and Cybernetics
(ICMLC), Vol. 3, pp- 1524 - 1527 ; 2010.
[17]. P. Somol, P. Pudil; “Feature Selection Toolbox”; Pattern recognition, Published by Elsevier Science
Ltd; 2002.
[18]. Pingfan Yan; Tong Chang ; “Best First Strategy for Feature Selection”; 9th
International Journal on
Pattern recognition, Vol.2, pp-706-708; 1988.
[19]. Manoranjan Dash, Huan Liu; “Consistency based feature selection”; Published by Elsevier
[20]. Computer Science (Artificial Intelligence); 2003.
[21]. http://www.di.unipi.it/~bacciu/teaching/IIA2012/lect3-exploratory-hand.pdf
[22]. T. Hastie, R. Tibshirani, and J. Friedman; “The Elements of Statistical Learning”; Springer, 2001.
[23]. Huan Liu and Lei Yu;” Toward Integrating Feature Selection Algorithms for Classification and
Clustering”; IEEE journal of Knowledge and data engineering; volume 17, issue 4; pp. 491-502; 2005
Shweta Srivastava, birth place is Faizabad and date of birth is 2-July-1986. She has done
schooling from Canossa Convent Girls Inter College, Faizabad, Uttar Pradesh, India, B.
Tech. (IT) from JSS Academy of Technical Education, NOIDA, Uttar Pradesh Technical
University, Uttar Pradesh, India (2008) and M. Tech. (CSE) from Jaypee Institute of
Information Technology, NOIDA, JIITU, Uttar Pradesh, India (2011). She has three years
of teaching experience. Currently she is working as an ASSISTANT PROFESSOR in
CSE department of ABES Engineering College, Ghaziabad, Uttar Pradesh, India. She has
worked as a LECTURER in JSSATE, NOIDA, Uttar Pradesh, India for one year. She has
done two years TEACHING ASSISTANTSHIP in Jaypee Institute of Information
Technology, JIITU, NOIDA, Uttar Pradesh, India during M. Tech.

More Related Content

PDF
Booster in High Dimensional Data Classification
PDF
New Feature Selection Model Based Ensemble Rule Classifiers Method for Datase...
PDF
A Threshold fuzzy entropy based feature selection method applied in various b...
PDF
Network Based Intrusion Detection System using Filter Based Feature Selection...
PDF
M43016571
PDF
Unsupervised Feature Selection Based on the Distribution of Features Attribut...
PDF
The International Journal of Engineering and Science (The IJES)
PDF
A novel hybrid feature selection approach
Booster in High Dimensional Data Classification
New Feature Selection Model Based Ensemble Rule Classifiers Method for Datase...
A Threshold fuzzy entropy based feature selection method applied in various b...
Network Based Intrusion Detection System using Filter Based Feature Selection...
M43016571
Unsupervised Feature Selection Based on the Distribution of Features Attribut...
The International Journal of Engineering and Science (The IJES)
A novel hybrid feature selection approach

What's hot (18)

PDF
AN EFFICIENT FEATURE SELECTION MODEL FOR IGBO TEXT
PDF
An unsupervised feature selection algorithm with feature ranking for maximizi...
PDF
An efficient feature selection in
PDF
G046024851
PDF
Optimization Technique for Feature Selection and Classification Using Support...
PDF
Feature Selection Algorithm for Supervised and Semisupervised Clustering
PDF
Filter Based Approach for Genomic Feature Set Selection (FBA-GFS)
PDF
Feature selection for classification
PDF
Improving the performance of Intrusion detection systems
DOC
Research proposal
PDF
A Survey and Comparative Study of Filter and Wrapper Feature Selection Techni...
PDF
A Survey on Constellation Based Attribute Selection Method for High Dimension...
PPTX
Research Proposal
DOCX
A fast clustering based feature subset selection algorithm for high-dimension...
DOCX
A fast clustering based feature subset selection algorithm for high-dimension...
PDF
EFFICIENT FEATURE SUBSET SELECTION MODEL FOR HIGH DIMENSIONAL DATA
PDF
F363941
PDF
Feature selection for multiple water quality status: integrated bootstrapping...
AN EFFICIENT FEATURE SELECTION MODEL FOR IGBO TEXT
An unsupervised feature selection algorithm with feature ranking for maximizi...
An efficient feature selection in
G046024851
Optimization Technique for Feature Selection and Classification Using Support...
Feature Selection Algorithm for Supervised and Semisupervised Clustering
Filter Based Approach for Genomic Feature Set Selection (FBA-GFS)
Feature selection for classification
Improving the performance of Intrusion detection systems
Research proposal
A Survey and Comparative Study of Filter and Wrapper Feature Selection Techni...
A Survey on Constellation Based Attribute Selection Method for High Dimension...
Research Proposal
A fast clustering based feature subset selection algorithm for high-dimension...
A fast clustering based feature subset selection algorithm for high-dimension...
EFFICIENT FEATURE SUBSET SELECTION MODEL FOR HIGH DIMENSIONAL DATA
F363941
Feature selection for multiple water quality status: integrated bootstrapping...
Ad

Similar to International Journal of Engineering Research and Development (IJERD) (20)

PDF
763354.MIPRO_2015_JovicBrkicBogunovic.pdf
PDF
Study on Relavance Feature Selection Methods
PDF
JUNE-77.pdf
PDF
ATTRIBUTE REDUCTION-BASED ENSEMBLE RULE CLASSIFIERS METHOD FOR DATASET CLASSI...
PDF
A Survey on Classification of Feature Selection Strategies
PDF
Effective Feature Selection for Feature Possessing Group Structure
PPT
feature selection slides share and types of features selection
PDF
Feature selection techniques for microarray dataset: a review
PDF
A Review on Feature Selection Methods For Classification Tasks
PDF
C LUSTERING B ASED A TTRIBUTE S UBSET S ELECTION U SING F AST A LGORITHm
PDF
AN EFFICIENT FEATURE SELECTION IN CLASSIFICATION OF AUDIO FILES
PPT
few common Feature of Size Datum Features are bores, cylinders, slots, or tab...
PDF
research paper
PPTX
Feature Engineering Fundamentals Explained.pptx
PDF
Survey on semi supervised classification methods and feature selection
PDF
Survey on semi supervised classification methods and
PPT
6811067.ppt6811067.ppt6811067.ppt6811067.ppt
PDF
Data Mining Module 2 Business Analytics.
PDF
Survey on software remodularization techniques
PDF
Survey on software remodularization techniques
763354.MIPRO_2015_JovicBrkicBogunovic.pdf
Study on Relavance Feature Selection Methods
JUNE-77.pdf
ATTRIBUTE REDUCTION-BASED ENSEMBLE RULE CLASSIFIERS METHOD FOR DATASET CLASSI...
A Survey on Classification of Feature Selection Strategies
Effective Feature Selection for Feature Possessing Group Structure
feature selection slides share and types of features selection
Feature selection techniques for microarray dataset: a review
A Review on Feature Selection Methods For Classification Tasks
C LUSTERING B ASED A TTRIBUTE S UBSET S ELECTION U SING F AST A LGORITHm
AN EFFICIENT FEATURE SELECTION IN CLASSIFICATION OF AUDIO FILES
few common Feature of Size Datum Features are bores, cylinders, slots, or tab...
research paper
Feature Engineering Fundamentals Explained.pptx
Survey on semi supervised classification methods and feature selection
Survey on semi supervised classification methods and
6811067.ppt6811067.ppt6811067.ppt6811067.ppt
Data Mining Module 2 Business Analytics.
Survey on software remodularization techniques
Survey on software remodularization techniques
Ad

More from IJERD Editor (20)

PDF
A Novel Method for Prevention of Bandwidth Distributed Denial of Service Attacks
PDF
MEMS MICROPHONE INTERFACE
PDF
Influence of tensile behaviour of slab on the structural Behaviour of shear c...
PDF
Gold prospecting using Remote Sensing ‘A case study of Sudan’
PDF
Reducing Corrosion Rate by Welding Design
PDF
Router 1X3 – RTL Design and Verification
PDF
Active Power Exchange in Distributed Power-Flow Controller (DPFC) At Third Ha...
PDF
Mitigation of Voltage Sag/Swell with Fuzzy Control Reduced Rating DVR
PDF
Study on the Fused Deposition Modelling In Additive Manufacturing
PDF
Spyware triggering system by particular string value
PDF
A Blind Steganalysis on JPEG Gray Level Image Based on Statistical Features a...
PDF
Secure Image Transmission for Cloud Storage System Using Hybrid Scheme
PDF
Application of Buckley-Leverett Equation in Modeling the Radius of Invasion i...
PDF
Gesture Gaming on the World Wide Web Using an Ordinary Web Camera
PDF
Hardware Analysis of Resonant Frequency Converter Using Isolated Circuits And...
PDF
Simulated Analysis of Resonant Frequency Converter Using Different Tank Circu...
PDF
Moon-bounce: A Boon for VHF Dxing
PDF
“MS-Extractor: An Innovative Approach to Extract Microsatellites on „Y‟ Chrom...
PDF
Importance of Measurements in Smart Grid
PDF
Study of Macro level Properties of SCC using GGBS and Lime stone powder
A Novel Method for Prevention of Bandwidth Distributed Denial of Service Attacks
MEMS MICROPHONE INTERFACE
Influence of tensile behaviour of slab on the structural Behaviour of shear c...
Gold prospecting using Remote Sensing ‘A case study of Sudan’
Reducing Corrosion Rate by Welding Design
Router 1X3 – RTL Design and Verification
Active Power Exchange in Distributed Power-Flow Controller (DPFC) At Third Ha...
Mitigation of Voltage Sag/Swell with Fuzzy Control Reduced Rating DVR
Study on the Fused Deposition Modelling In Additive Manufacturing
Spyware triggering system by particular string value
A Blind Steganalysis on JPEG Gray Level Image Based on Statistical Features a...
Secure Image Transmission for Cloud Storage System Using Hybrid Scheme
Application of Buckley-Leverett Equation in Modeling the Radius of Invasion i...
Gesture Gaming on the World Wide Web Using an Ordinary Web Camera
Hardware Analysis of Resonant Frequency Converter Using Isolated Circuits And...
Simulated Analysis of Resonant Frequency Converter Using Different Tank Circu...
Moon-bounce: A Boon for VHF Dxing
“MS-Extractor: An Innovative Approach to Extract Microsatellites on „Y‟ Chrom...
Importance of Measurements in Smart Grid
Study of Macro level Properties of SCC using GGBS and Lime stone powder

Recently uploaded (20)

PDF
Build a system with the filesystem maintained by OSTree @ COSCUP 2025
PPTX
Cloud computing and distributed systems.
PDF
Unlocking AI with Model Context Protocol (MCP)
PDF
Modernizing your data center with Dell and AMD
PDF
Per capita expenditure prediction using model stacking based on satellite ima...
PDF
Building Integrated photovoltaic BIPV_UPV.pdf
PDF
Spectral efficient network and resource selection model in 5G networks
PDF
Advanced methodologies resolving dimensionality complications for autism neur...
PDF
How UI/UX Design Impacts User Retention in Mobile Apps.pdf
PDF
Review of recent advances in non-invasive hemoglobin estimation
PDF
Dropbox Q2 2025 Financial Results & Investor Presentation
PDF
Peak of Data & AI Encore- AI for Metadata and Smarter Workflows
PDF
Electronic commerce courselecture one. Pdf
PPTX
20250228 LYD VKU AI Blended-Learning.pptx
PDF
Chapter 3 Spatial Domain Image Processing.pdf
PDF
Network Security Unit 5.pdf for BCA BBA.
PDF
The Rise and Fall of 3GPP – Time for a Sabbatical?
PDF
Machine learning based COVID-19 study performance prediction
PDF
cuic standard and advanced reporting.pdf
DOCX
The AUB Centre for AI in Media Proposal.docx
Build a system with the filesystem maintained by OSTree @ COSCUP 2025
Cloud computing and distributed systems.
Unlocking AI with Model Context Protocol (MCP)
Modernizing your data center with Dell and AMD
Per capita expenditure prediction using model stacking based on satellite ima...
Building Integrated photovoltaic BIPV_UPV.pdf
Spectral efficient network and resource selection model in 5G networks
Advanced methodologies resolving dimensionality complications for autism neur...
How UI/UX Design Impacts User Retention in Mobile Apps.pdf
Review of recent advances in non-invasive hemoglobin estimation
Dropbox Q2 2025 Financial Results & Investor Presentation
Peak of Data & AI Encore- AI for Metadata and Smarter Workflows
Electronic commerce courselecture one. Pdf
20250228 LYD VKU AI Blended-Learning.pptx
Chapter 3 Spatial Domain Image Processing.pdf
Network Security Unit 5.pdf for BCA BBA.
The Rise and Fall of 3GPP – Time for a Sabbatical?
Machine learning based COVID-19 study performance prediction
cuic standard and advanced reporting.pdf
The AUB Centre for AI in Media Proposal.docx

International Journal of Engineering Research and Development (IJERD)

  • 1. International Journal of Engineering Research and Development e-ISSN: 2278-067X, p-ISSN: 2278-800X, www.ijerd.com Volume 7, Issue 6 (June 2013), PP. 57-61 57 A Review Paper on Feature Selection Methodologies and Their Applications Ms. Shweta Srivastava1 , Ms. Nikita Joshi2 , Ms. Madhvi Gaur3 Assistant Professor(CSE Department) ABES Engineering College, Ghaziabad, India. Abstract:- Feature selection is the process of eliminating features from the data set that are irrelevant with respect to the task to be performed. Feature selection is important for many reasons such as simplification, performance, computational efficiency and feature interpretability. It can be applied to both supervised and unsupervised learning methodologies. Such techniques are able in improving the efficiency of various machine learning algorithms and that of training as well. Feature selection speed up the run time of learning, improves data quality and data understanding. Keywords:- Feature Selection, Supervised, Search strategies, Unsupervised. I. INTRODUCTION Feature selection is used to satisfy the common goal of maximizing the accuracy of the classifier, minimizing the related measurement costs; improve accuracy by reducing irrelevant and possibly redundant features; reduce the complexity and the associated computational cost; and improve the probability that a solution will be comprehensible and realistic [13]. Feature selection is one of the stage for preprocessing the data to reduce the dimensionality [10]. It selects a subset of the existing features without any transformation. It can be said as a special case of feature extraction. II. FEATURE SELECTION METHODOLOGIES When we use feature selection, smaller number of features are extracted which means fewer model parameters. It improves the generalization capabilities and reduces complexity and execution time. Feature Selection methodologies can be categorized as: supervised and unsupervised feature selection methods. Supervised Feature Selection Method: A feature selection method requires a search strategy to select the subset of attributes and an objective function to evaluate the selected subset of features. Objective function is divided in three categories: 1. Filters: It consists of algorithms that are built in the adaptive systems for data analysis (predictors) [9]. They use an evaluation function that relies on properties of the data. Distance based and margin based criterion can be used for filters. Fig.1. Filter method 2. Wrapper: The algorithms of this method are wrapped around the adaptive systems providing them subsets of features and receiving their feedback (usually accuracy). These wrapper approaches are aimed at improving
  • 2. A Review Paper on Feature Selection Methodologies and Their Applications 58 results of the specific predictors they work with [9]. It utilizes the classifier as black box to find the subset of features based on their predictive power [14]. 3. Embedded: They are based on performance evaluation metric calculated directly from the data, without direct reference to the results of any data analysis systems. Fig.2. Wrapper method 3. Embedded: They are based on performance evaluation metric calculated directly from the data, without direct reference to the results of any data analysis systems. Filter Methods Wrapper methods Embedded methods Filter methods appear to be a probably less optimal.[4] Wrapper methods are superior alternative in supervised learning problems. [4] The performance of embedded method degrades if more irrelevant features are inserted in target set. Filter methods execute faster than wrapper methods. Wrapper methods execute slower than filter methods. Embedded method is faster than the wrapper methods. The result of filter method exhibits more generality than that of wrapper approach. There is lack of generality in wrapper methods as it is tied to some classifier. It also lacks generality as it is also dependent on some classification algorithm. Filter method has more tendencies to select large subset of data. Wrapper method is more accurate than filter methods as it achieves better recognition rates rather than that of filters. They are least prone to over fitting. Computational cost is less for large data set in filter method. Computational cost is more for large data set in filter method. Computational cost is less in comparison to wrapper methods. Independent of classification algorithm. Dependent on classification algorithm. Dependent on classification algorithm. Table 1: Comparison of filter, wrapper and embedded methods Search Strategy: A search strategy is required to select the candidate subsets and objective function evaluates each of the candidate subset. Search strategy involves exhaustive, heuristic and randomized searching algorithms [15][16]. The time complexity is exponential in terms of dimensionality for exhaustive search and quadratic for heuristic search. The complexity can be linear to the number of iterations in a random search [15]. There are several hybrid algorithms also. Exhaustive search: It evaluates a number of subsets that grows exponentially with the dimensionality of the search space. It finds the 2N combinations of all m features. Each of the subset is evaluated by objective function and a measure of goodness is returned to the search algorithm. Branch and bound based algorithms are used for
  • 3. A Review Paper on Feature Selection Methodologies and Their Applications 59 exhaustive search. There are several versions of branch and bound algorithm i.e. Basic BB (slowest), Enhanced BB (fastest), Fast BB, BB with partial prediction [17]. Heuristic search: The number of subsets evaluated by BFF is less (even much less) than that needed by the branch and bound algorithm [18]. They require large computational time. Heuristic search is problematic when the data has highly correlated features. Heuristics help to reduce the number of alternatives from an exponential number to a polynomial number [19]. For example: Sequential forward selection, sequential backward elimination and bidirectional search. Forward Selection: Forward selection considers the subset to be empty initially and keeps on adding one feature at a time until the best feature subset is obtained. Backward Elimination: Backward selection takes complete set of features as input and keeps on removing one attribute at a time until the most appropriate subset of features is obtained. Randomized search strategy: It performs randomized exploration of the search space where next direction is a sample from a given probability [20]. For example: genetic algorithm. Accuracy Complexity Advantages Disadvantages Exhaustive Search It always finds the best possible solution. Exponential It is highly accurate. Complexity is high. Heuristic Search This technique is good if no backtracking is needed. Quadratic It is simple and fast. Backtracking is not possible. Randomized Search It is good with proper control parameters. Generally low It is designed to escape local minima. It is difficult to choose good parameters. Table 2: Comparison of Exhaustive, Heuristic and Randomized Search Unsupervised feature selection: The procedure of arranging the objects into natural classes whose members are similar to each other, identified by a given metrics. Unsupervised feature selection is particularly difficult due to the absence of class labels for feature relevance estimation. Unsupervised feature selection is a less constrained search problem without class labels, depending on clustering quality measures and can achieve many equally valid feature subsets. Feature redundancy and relevance is measured with respect to clusters instead of classes. The objective of clustering is to maximize intra cluster similarity and minimizing inter cluster similarity. For example: Agglomerative and partitioned clustering. Filter and wrapper approach are helpful in unsupervised feature selection also. III. APPLICATION OF FEATURE SELECTION IN VARIOUS FIELDS Text classification: There are various challenges related to automated text classification such as: 1. An appropriate data structure is to be selected to represent the documents. 2. An appropriate objective function is to be chosen to optimize to avoid overfitting and obtain good generalization. Along with it algorithmic issues arising as a result of the high formal dimensionality of the data are to be dealt with [1]. Genre classification: Metadata such as filename, author, size, date, track length and genres are the common features used to classify and retrieve genre documents. On the basis of these data, the classification is infeasible, so the feature selection step is required. In case of genre classification feature selection is a process where a segment of an audio is characterized into a compact numerical representation [2]. Feature selection is done to reduce the dimensionality of the data as a preprocessing step prior to classification due to high dimensionality of the feature sets. Microarray data analysis: 1. Almost all bioinformatics problems have the number of features significantly larger than the number of samples (high feature to sample ratio datasets) [3]. For example: Breast cancer classification on the basis of microarray data. Though the information about all the genes is not required in case of such classification. 2. Content analysis and signal analysis in genomics also require feature selection.
  • 4. A Review Paper on Feature Selection Methodologies and Their Applications 60 Software defect prediction: There are various software quality assurance attributes such as reliability, functionality, fault proneness, reusability, comprehensibility etc [6]. It is a critical issue to select most appropriate software metrics that likely to indicate fault proneness. Sentiment analysis: Sentiment analysis is capturing favorability using natural language processing. It is not just a topic based categorization. It deals with the computational treatment of opinion, sentiment, and subjectivity in text. It is useful in recommendation systems and question answering [8]. To decide about the positivity or negativity of the opinion on the basis of various features such as term presence, feature frequency, feature presence, term position, POS tags, syntax, topic and negation etc. All of the features are not required in each and every case. So feature selection need to be performed. Stock market analysis: There are hundreds of stock index futures. Along with it financial data including the stock market data is too complex to be searched easily [11]. In particular, the existence of large amount of continuous data may cause a challenging task to explicit concepts extraction from the raw data due to the huge amount of data space determined by continuous features [12]. So it is necessary to reduce the dimensionality of data and irrelevant factors before searching. Image Retrieval Feature selection is applied to content based image retrieval to allow efficient browsing, searching and retrieving [22]. Content based image retrieval is to index the images on the basis of their own visual contents (i.e. color, shape, texture etc.) instead of text based keyword indexing. The biggest problem for content based image retrieval is large amount of images in database [21]. IV. CONCLUSION This paper provides a comprehensive overview of various characteristic of feature selection. Feature selection as a preprocessing step in very large databases collected from various applications. More work is required to overcome limitations imposed as it is costly to visit high dimensionality data multiple times or accessing instances at random times. In unsupervised feature selection, several clusters may exist in different subspaces of small-small dimensionality, with their sets of dimensions overlapped or non-overlapped. As most existing feature selection algorithms have quadratic or higher time complexity about N, so it is quite difficult to handle high dimensionality. Therefore, more efficient search strategies and evaluation criteria are needed for feature selection with large dimensionality. REFERENCES [1]. Anirban Dasgupta,Petros Drineas, Boulos Harb, Vanja Josifovski, Michael W. Mahoney; “Feature Selection Methods for Text Classification”,KDD- ACM.,2007. [2]. Shyamala Doraisamy, Shahram Golzari, Noris Mohd. Norowi, Md. Nasir B Sulaiman, Nur Izura Udzir; ” A Study on Feature Selection and Classification Techniques for Automatic Genre Classification of Traditional Malay Music” Proceedings of ISMIR, 2008. [3]. Gianluca Bontempi, Benjamin Haibe-Kains; “Feature selection methods for mining bioinformatics data”, [4]. Luis Talavera; “An evaluation of filter and wrapper methods for feature selection in categorical clustering”, Proceedings of the 6th international conference on Advances in Intelligent Data Analysis, 2005. [5]. Yvan Saeys, Inaki Inza, Pedro Larranaga; ”A review of feature selection techniques in bioinformatics”, Oxford Journals, 2007. [6]. N.Gayatri, S.Nickolas, A.V.Reddy; “Feature Selection Using Decision Tree Induction in Class level Metrics Dataset for Software Defect Predictions”; Proceedings of the World Congress on Engineering and Computer Science ; Vol I; 2010. [7]. Tim O’Keefe, Irena Koprinska;” Feature Selection andWeighting Methods in Sentiment Analysis”; Proceedings of the 14th Australasian Document Computing Symposium, Sydney, Australia; 2009. [8]. Bo Pang, Lillian Lee; ”Opinion Mining and Sentiment Analysis”; Foundations and Trends in Information Retrieval, Vol. 2, Nos. 1–2 pp- 1–135, 2008. [9]. Wlodzislaw Duch; “Filter methods”, Springer- Feature Extraction Studies in Fuzziness and Soft Computing Volume 207, 2006, pp 89-117.
  • 5. A Review Paper on Feature Selection Methodologies and Their Applications 61 [10]. Chih- Fong Tsai; “Data pre-processing by genetic algorithms for bankruptcy prediction”, IEEE International Conference on Industrial Engineering and Engineering Management, Pages- 1780 - 1783 , 2011. [11]. Kyoung-jae Kim, Ingoo Han; “Genetic algorithms approach to feature discretization in artificial neural networks for the prediction of stock price index”; Expert Systems with Applications, Elsevier Science Ltd; 2000. [12]. Liu, H., & Setiono, R.; “Dimensionality reduction via discretization”; Knowledge-Based Systems, 9 (1), 67–72; 1996. [13]. Steppe. J., K.W. Bauer, “Feature Saliency Measures”, Computers & Mathematics with Applications, Vol 33, No. 8, pp. 109-126; 1997. [14]. L.Ladha et al., “Feature Selection Methods and Algorithms”; International Journal on Computer Science and Engineering (IJCSE), Vol. 3. No. 5, pp. 1787-1797; 2011. [15]. Roberto Ruiz, Jos´e C. Riquelme, and Jes´us S. Aguilar-Ruiz; “Heuristic Search over a Ranking for Feature Selection”; IWANN, LNCS 3512, pp. 742–749; 2005. [16]. Yao-Hong Chan; “Empirical comparison of forward and backward search strategies in L-GEM based feature selection with RBFNN”; International Conference on Machine Learning and Cybernetics (ICMLC), Vol. 3, pp- 1524 - 1527 ; 2010. [17]. P. Somol, P. Pudil; “Feature Selection Toolbox”; Pattern recognition, Published by Elsevier Science Ltd; 2002. [18]. Pingfan Yan; Tong Chang ; “Best First Strategy for Feature Selection”; 9th International Journal on Pattern recognition, Vol.2, pp-706-708; 1988. [19]. Manoranjan Dash, Huan Liu; “Consistency based feature selection”; Published by Elsevier [20]. Computer Science (Artificial Intelligence); 2003. [21]. http://www.di.unipi.it/~bacciu/teaching/IIA2012/lect3-exploratory-hand.pdf [22]. T. Hastie, R. Tibshirani, and J. Friedman; “The Elements of Statistical Learning”; Springer, 2001. [23]. Huan Liu and Lei Yu;” Toward Integrating Feature Selection Algorithms for Classification and Clustering”; IEEE journal of Knowledge and data engineering; volume 17, issue 4; pp. 491-502; 2005 Shweta Srivastava, birth place is Faizabad and date of birth is 2-July-1986. She has done schooling from Canossa Convent Girls Inter College, Faizabad, Uttar Pradesh, India, B. Tech. (IT) from JSS Academy of Technical Education, NOIDA, Uttar Pradesh Technical University, Uttar Pradesh, India (2008) and M. Tech. (CSE) from Jaypee Institute of Information Technology, NOIDA, JIITU, Uttar Pradesh, India (2011). She has three years of teaching experience. Currently she is working as an ASSISTANT PROFESSOR in CSE department of ABES Engineering College, Ghaziabad, Uttar Pradesh, India. She has worked as a LECTURER in JSSATE, NOIDA, Uttar Pradesh, India for one year. She has done two years TEACHING ASSISTANTSHIP in Jaypee Institute of Information Technology, JIITU, NOIDA, Uttar Pradesh, India during M. Tech.