SlideShare a Scribd company logo
International Journal on Soft Computing (IJSC) Vol.6, No. 1, February 2015
DOI: 10.5121/ijsc.2015.6105 53
Classification of Vehicles Based on Audio Signals
using Quadratic Discriminant Analysis and High
Energy Feature Vectors
A. D. Mayvana
, S. A. Beheshtib
, M. H. Masoomc
a
Department of Electrical Engineering, Iran University of Science and
Technology, Tehran, Iran
b
Department of Electrical Engineering, Iran University of Science and
Technology, Tehran, Iran.
c
Department of Mechanical Engineering, BabolNoshirvani University of
Technology, Babol, Iran.
ABSTRACT
The focusof this paper is on classification of different vehicles using sound emanated from the vehicles. In
this paper,quadratic discriminant analysis classifies audio signals of passing vehicles to bus, car, motor, and
truck categories based on features such as short time energy, average zero cross rate, and pitch frequency of
periodic segments of signals. Simulation results show that just by considering high energy feature vectors,
better classification accuracy can be achieved due to the correspondence of low energy regions with noises
of the background. To separate these elements, short time energy and average zero cross rate are used
simultaneously.In our method,we have used a few features which are easy to be calculated in time domain
and enable practical implementation of efficient classifier. Although, the computation complexity is low,
the classification accuracy is comparable with other classification methodsbased on long feature vectors
reported in literature for this problem.
KEYWORD
Classification accuracy; Periodic segments; Quadratic Discriminant Analysis; Separation criterion; Short
time analysis.
1. INTRODUCTION
Vehicle identification while it is in motion is a prerequisite for traffic and speed management,
classified vehicle count, traffic signal time optimization, gap/ headway measurement, and military
purposes. Moving vehicles affect the environment in different ways. Vehicle emits heats, sounds,
and magnetic field. There are many approaches investigated vehicle identification based on
different kinds of signals. Image processing techniques are used to classify vehicles under the real
time traffic management and for Intelligent Transportation Systems (ITS). Inductive loops based
International Journal on Soft Computing (IJSC) Vol.6, No. 1, February 2015
54
systems are widely used for determination of vehicle counts [1]. The most promising approach
for vehicle identification is the one that is based on acoustic signals. Moving vehicles emit
characteristic sounds. These sounds are generated from moving parts, frictions, winds, emissions,
tires,etc. Assuming that similar vehicles which have the same working conditions generate the
same sounds; then these sounds can be used to classify vehicles [2]. Vehicle classification based
on sound signals have already attempted by researchers. Paulrajet al. (2013) [3] used
autoregressive modeling algorithm for the analysis to extract the features from the recorded
vehicle signals. Probabilistic neural network (PNN) models are developed to classify the vehicle
type and its distance. Nooralahiyan et al. (1997) [4] used a directional microphone connected to a
DAT (Digital Audio Tape) recorder. The digital signal was pre-processed by LPC (Linear
Predictive Coding) parameter conversion based on autocorrelation analysis. A Time Delay Neural
Network (TDNN) was chosen to classify individual travelling vehicles based ontheir speed
independent acoustic signature to four broad categories: buses or Lorries, small or large saloons,
various types of motorcycles, and light goods vehicles or vans. Ghiurcau and Rusu(2009) [5]
presented a vehicle sound classification system using time encoded signal processing and
recognition (TESPAR) method combined with the archetypes technique implemented on the
Matlab platform. The experimental results show the efficiency of the TESPAR method when
dealing with vehicle sounds. Effectiveness of features set is usually measured by how well they
represent the signal; however, features should also satisfy several conditions: their number should
be small (less than 15); also,to enable practical implementation they should be easy to calculate.
There is a number of feature extraction methods used in vehicle classification. Some of them
produce too many features for a single input vector like estimation of Power Spectrum Density
[6], [7], [8], and some are too complicated like Principal Component Analysis [9], [7].
Nevertheless, no feature extraction method fulfills all our requirements, and there is no systematic
way to select best feature according to our criterion. Harmonic line, Schur coefficients, and MEL
filters, were presented in [10] each relating to properties of vehicle audio. Methods were
compared in context of their separability and correct classification rate. Aljaafreh and Dong
(2010) [2] investigated two feature extraction methods for acoustic signals from moving vehicles.
The first one is based on spectrum distribution and the second one on wavelet packet transform.
They evaluated the performance of different classifiers such as K-nearest neighbor algorithm
(KNN) and support vector machine (SVM). It is found that for vehicle sound data, a discrete
spectrum based feature extraction method outperforms wavelet packet transform method.
Experimental results verified that support vector machine is an efficient classifier for
vehicles.Other spectrum based feature extraction can be found in [11], [12], [13], [14], and [15].
Similarly, Discrete Wavelet Transform (DWT) is used in [16] and [17] to extract features using
statistical parameters and energy content of the wavelet coefficients.
This paper describes an algorithm to classify audio signals of vehicles. Quadratic discriminant
analysis uses feature vectors of periodic segments with elements: short time energy, average zero
cross rate, and pitch frequency to distinguish between vehicles’ signals. If we separate feature
vectors with high energy from others properly and train classifier with these vectors, better
classification accuracy will be achieved due to the correspondence of low energy regions and
noises of the background, as simulation results confirm. Obviously high energy feature vectors
have less zero cross rate. Considering this point is the basic idea for separation criterion. Our
innovation in separation is assuming vectors by energy larger than [], average of all
segments’ energy, and ZCR less than [],average of all segments’ ZCR, are high energy
elements. We have used three features which are easy to be calculated in time domain, which
enable practical implementation of efficient classifier. The cost paid is lower, but still satisfactory,
International Journal on Soft Computing (IJSC) Vol.6, No. 1, February 2015
55
classification accuracy than other classification applications reported in literature for this
problem. Sections of paper are as follow: After introduction first,we determine important
parameters and features.Then we study quadratic discriminant analysis and high energy element
separation criterion in section III. Simulation results show improvement in classification accuracy
for separation criterion. Finally, the paper ends by conclusion.
2. FEATURE EXTRACTION
Changing the characteristics of an audio signal over time seems normal. For example, Changes in
the signal’s peak value or rate of sign changes of amplitude are such a time varying
features.Although these changes exist in signals, short time analysis is a method to illustrate them
asobvious way. Basic assumption of this method is based on these changes occur slowly; then
wedivide signal into smaller parts and features are extracted.
2.1. Short Time Energy
This feature indicates how the signal amplitude changes over the time which is defined as
follows:
 = ∑ [
− 
]
∞
∞ (1)
Where 	
is the audio signal and
is the window that slides along the audio sequence
selecting the interval to be involved in the computation.
2.2. Short Time Average Zero Cross Rate
This feature indicates how the signal sign changes over the time which is defined as follows:
 = ∑ |
[	
] − 
[	
 − 1]|
∞
∞
− 
 (2)
Where
= 1 2
⁄ 												0 ≤ 
 ≤  − 1
																																				= 0																ℎ !
(3)
The rate at which signal sign changes occur is a simple measure of the frequency content of a
signal. This is particularly true for narrowband signals. For example, a sinusoidal signal of
frequency #$, sampled at a rate#%, has 2 #$ #%
⁄ average rate of zero-crossings.Thus, the average
ZCR gives a resonable way to estimate the frequency of a sine wave. Audio signals are
broadband signals and the interpretation of average ZCR is therefore much less precise. However,
rough estimates of spectral properties can be obtained using a represantation based onthe short
time energy and zero cross rate together.This representation was proposed by Reddy, and studied
by Vicens [18] as the basis for a large-scale speech recognition system.
International Journal on Soft Computing (IJSC) Vol.6, No. 1, February 2015
56
2.3. Pitch Frequency
Pitch period estimation (or fundamental frequency) is one of the most important problems in
speech processing. Pitch detectors are used in vocoders [19], speaker identification and
verification systems [20, 21]. Because of its importance, many solutions to this problem have
been proposed [22]. In this paper we use the fact thatthe autocorrelation function for periodic
segments attains a maximum at samples 0, ±',±2',… where T is pitch period. Since, there may
be exist other autocorrelation peaks except those due to the periodicity, the simple procedure of
picking the largest peak at sample'in the autocorrelation will failto estimate Pitch period. To
avoid this problem it is useful to process the signal so as to make periodicity more prominent
while suppressing other distracting features of the signal. Spectrum flattener with the objective to
bring each harmonic to the same amplitude level as in the case of a periodic impulse trainhas
been used. Specifically, center clipped signalhas been used in computing autocorrelation function.
In the scheme proposed,similar to [23], the center clipped signal is obtained by following
nonlinear transformation
For samples above (), the output is equal to the input minus the clipping level. For samples below
the clipping level the output is zero.We set clipping level, (), equal to 68% of the minimum of
two maximum amplitudes found in both the first third and last third of the audiosegment;denoted
respectively by
*+ and 
*. However, for the un-periodic segment, there are no strong
autocorrelation periodicity peaks and it seems to be a high frequency noise-like waveform [22].
We assume segments with weak peaks below 30 % clipped signal’s energy, ,)-.
=
∑ [,)-.
− 
]
∞
∞ ,are un-periodic. Finally pitch frequencies of frames smoothed
through median operation. Following figures present features.
−()()
Fig.1: center clipping function
International Journal on Soft Computing (IJSC) Vol.6, No. 1, February 2015
57
Fig.2: Energy for Bus in first row for three differentsounds, for Car in the second row, for Motor in the
third row, and for Truck in the fourth row.
Fig.3:ZCR for Bus in first row for three different sounds, for Car in the second row, for Motor in the third
row, and for Truck in the fourth row.
International Journal on Soft Computing (IJSC) Vol.6, No. 1, February 2015
58
Fig.4: Pitch Frequency for Bus in first row for three different sounds, for Car in the second row, for Motor
in the third row, and for Truck in the fourth row.
Features are extracted from rectangular windowed segment of length 15 Milliseconds and
overlapping of 5 Milliseconds between consecutive windows. Feature vectors of periodic
segments are used for training and classification. After determining the important features and
parameters for audio signal modeling, we will review quadratic discriminant analysis and high
energy separation criterion will be described.
3. CLASSIFIER AND SEPARATION CRITERION
Now the discriminant boundaries between classes should be determined. Quadratic
discriminantanalysis is used to discriminate the classes. The classifier that separate K classes has
K quadratic functions as follow [24]:
/0	1 = 12
31 + 52
1 + 6$ (4)
The coefficient6$is constant coefficient,5is vector of linear coefficients, and 3 is matrix
containing quadratic coefficients. Feature vectorX will belong to class k, if/0	1  /8	1;	∀; ≠
=.In classification step, after feature vectors were extracted, we are able to define class of each
vector separately. Finally, signal will allocate to a classwhichachieved great supply in feature
vector classification phase. In other words, probability of belonging to each class is proportional
to abundance of that class in allocated labels to vectors by quadratic discriminant analysis.

More Related Content

PDF
Classification of vehicles based on audio signals
PDF
Jamming Detection based on Doppler Shift Estimation in Vehicular Communicatio...
PDF
Enhancing Performance for Orthogonal Frequency Division Multiplexing in Wirel...
PDF
Performance analysis of adaptive filter channel estimated MIMO OFDM communica...
PDF
AN MINIMUM RECONFIGURATION PROBABILITY ROUTING ALGORITHM FOR RWA IN ALL-OPTIC...
PDF
Topology estimation of a digital subscriber line
PDF
Signal classification of second order cyclostationarity signals using bt scld...
PDF
Routing in All-Optical Networks Using Recursive State Space Technique
Classification of vehicles based on audio signals
Jamming Detection based on Doppler Shift Estimation in Vehicular Communicatio...
Enhancing Performance for Orthogonal Frequency Division Multiplexing in Wirel...
Performance analysis of adaptive filter channel estimated MIMO OFDM communica...
AN MINIMUM RECONFIGURATION PROBABILITY ROUTING ALGORITHM FOR RWA IN ALL-OPTIC...
Topology estimation of a digital subscriber line
Signal classification of second order cyclostationarity signals using bt scld...
Routing in All-Optical Networks Using Recursive State Space Technique

What's hot (19)

PDF
DORA: Server Based VANETs and its Applications
PDF
Analyzing the performance of the dynamic
PDF
40120140504004
PDF
A Diagnostic Analytics of Harmonic Source Signature Recognition by Using Peri...
PDF
An Improved of Multiple Harmonic Sources Identification in Distribution Syste...
PDF
Ijetcas14 344
PDF
Qo s parameters for obs network
PDF
Performance Analysis and Comparative Study of Cognitive Radio Spectrum Sensin...
PDF
K017426872
PDF
OPTIMIZATION OF THE RECURSIVE ONE-SIDED HYPOTHESIS TESTING TECHNIQUE FOR AUTO...
PPTX
PDF
IRJET-Study of Performance analysis of Wind Tunnel Simulation of Pollutant Di...
PDF
HIGH RESOLUTION METHOD FOR DOA ESTIMATION
PDF
A Summative Comparison of Blind Channel Estimation Techniques for Orthogonal ...
PDF
A COMPARISON BETWEEN SWARM INTELLIGENCE ALGORITHMS FOR ROUTING PROBLEMS
PDF
Travel Time Prediction Using Dedicated Short-Range Communications Probe Data
PPTX
Fakhre alam
PDF
IRJET- Channel and Rate Selection in Cognitive Radio Network
DORA: Server Based VANETs and its Applications
Analyzing the performance of the dynamic
40120140504004
A Diagnostic Analytics of Harmonic Source Signature Recognition by Using Peri...
An Improved of Multiple Harmonic Sources Identification in Distribution Syste...
Ijetcas14 344
Qo s parameters for obs network
Performance Analysis and Comparative Study of Cognitive Radio Spectrum Sensin...
K017426872
OPTIMIZATION OF THE RECURSIVE ONE-SIDED HYPOTHESIS TESTING TECHNIQUE FOR AUTO...
IRJET-Study of Performance analysis of Wind Tunnel Simulation of Pollutant Di...
HIGH RESOLUTION METHOD FOR DOA ESTIMATION
A Summative Comparison of Blind Channel Estimation Techniques for Orthogonal ...
A COMPARISON BETWEEN SWARM INTELLIGENCE ALGORITHMS FOR ROUTING PROBLEMS
Travel Time Prediction Using Dedicated Short-Range Communications Probe Data
Fakhre alam
IRJET- Channel and Rate Selection in Cognitive Radio Network
Ad

Similar to Classification of Vehicles Based on Audio Signals using Quadratic Discriminant Analysis and High Energy Feature Vectors (20)

PDF
March 2021: Top Read Articles in Soft Computing
PDF
July 2025: Top Read Articles in Soft Computing
PDF
Exploring Sound Signature for Vehicle Detection and Classification Using ANN
PDF
December 2024 - Top Read Articles in Soft Computing
PDF
October 2021: Top Read Articles in Soft Computing
PDF
EXPLORING SOUND SIGNATURE FOR VEHICLE DETECTION AND CLASSIFICATION USING ANN
PDF
EXPLORING SOUND SIGNATURE FOR VEHICLE DETECTION AND CLASSIFICATION USING ANN
PDF
Multi-Target Classification Using Acoustic Signatures in Wireless Sensor Netw...
PDF
Fault detection of motorcycles using the Slopes of the estimated pseudospectr...
PDF
Vehicle Noise Pattern Recognition by Self-Organizing Maps
PDF
A computationally efficient learning model to classify audio signal attributes
PDF
IRJET- Music Genre Classification using SVM
PDF
Broad phoneme classification using signal based features
PDF
T26123129
PDF
Ijarcet vol-2-issue-4-1347-1351
PDF
Investigating Multi-Feature Selection and Ensembling for Audio Classification
PDF
Broad Phoneme Classification Using Signal Based Features
PDF
IRJET- Musical Instrument Recognition using CNN and SVM
PDF
IRJET- Music Genre Classification using GMM
DOCX
Optimized audio classification and segmentation algorithm by using ensemble m...
March 2021: Top Read Articles in Soft Computing
July 2025: Top Read Articles in Soft Computing
Exploring Sound Signature for Vehicle Detection and Classification Using ANN
December 2024 - Top Read Articles in Soft Computing
October 2021: Top Read Articles in Soft Computing
EXPLORING SOUND SIGNATURE FOR VEHICLE DETECTION AND CLASSIFICATION USING ANN
EXPLORING SOUND SIGNATURE FOR VEHICLE DETECTION AND CLASSIFICATION USING ANN
Multi-Target Classification Using Acoustic Signatures in Wireless Sensor Netw...
Fault detection of motorcycles using the Slopes of the estimated pseudospectr...
Vehicle Noise Pattern Recognition by Self-Organizing Maps
A computationally efficient learning model to classify audio signal attributes
IRJET- Music Genre Classification using SVM
Broad phoneme classification using signal based features
T26123129
Ijarcet vol-2-issue-4-1347-1351
Investigating Multi-Feature Selection and Ensembling for Audio Classification
Broad Phoneme Classification Using Signal Based Features
IRJET- Musical Instrument Recognition using CNN and SVM
IRJET- Music Genre Classification using GMM
Optimized audio classification and segmentation algorithm by using ensemble m...
Ad

Recently uploaded (20)

PDF
July 2025 - Top 10 Read Articles in International Journal of Software Enginee...
PPTX
CH1 Production IntroductoryConcepts.pptx
PPTX
Lecture Notes Electrical Wiring System Components
PPTX
IOT PPTs Week 10 Lecture Material.pptx of NPTEL Smart Cities contd
PDF
The CXO Playbook 2025 – Future-Ready Strategies for C-Suite Leaders Cerebrai...
PDF
Well-logging-methods_new................
PDF
Embodied AI: Ushering in the Next Era of Intelligent Systems
PDF
R24 SURVEYING LAB MANUAL for civil enggi
PPT
Mechanical Engineering MATERIALS Selection
PPTX
Welding lecture in detail for understanding
PDF
SM_6th-Sem__Cse_Internet-of-Things.pdf IOT
PPTX
Sustainable Sites - Green Building Construction
PDF
TFEC-4-2020-Design-Guide-for-Timber-Roof-Trusses.pdf
PPTX
additive manufacturing of ss316l using mig welding
PPTX
Engineering Ethics, Safety and Environment [Autosaved] (1).pptx
PDF
Automation-in-Manufacturing-Chapter-Introduction.pdf
PPT
CRASH COURSE IN ALTERNATIVE PLUMBING CLASS
PDF
Operating System & Kernel Study Guide-1 - converted.pdf
PPTX
Geodesy 1.pptx...............................................
PPTX
bas. eng. economics group 4 presentation 1.pptx
July 2025 - Top 10 Read Articles in International Journal of Software Enginee...
CH1 Production IntroductoryConcepts.pptx
Lecture Notes Electrical Wiring System Components
IOT PPTs Week 10 Lecture Material.pptx of NPTEL Smart Cities contd
The CXO Playbook 2025 – Future-Ready Strategies for C-Suite Leaders Cerebrai...
Well-logging-methods_new................
Embodied AI: Ushering in the Next Era of Intelligent Systems
R24 SURVEYING LAB MANUAL for civil enggi
Mechanical Engineering MATERIALS Selection
Welding lecture in detail for understanding
SM_6th-Sem__Cse_Internet-of-Things.pdf IOT
Sustainable Sites - Green Building Construction
TFEC-4-2020-Design-Guide-for-Timber-Roof-Trusses.pdf
additive manufacturing of ss316l using mig welding
Engineering Ethics, Safety and Environment [Autosaved] (1).pptx
Automation-in-Manufacturing-Chapter-Introduction.pdf
CRASH COURSE IN ALTERNATIVE PLUMBING CLASS
Operating System & Kernel Study Guide-1 - converted.pdf
Geodesy 1.pptx...............................................
bas. eng. economics group 4 presentation 1.pptx

Classification of Vehicles Based on Audio Signals using Quadratic Discriminant Analysis and High Energy Feature Vectors

  • 1. International Journal on Soft Computing (IJSC) Vol.6, No. 1, February 2015 DOI: 10.5121/ijsc.2015.6105 53 Classification of Vehicles Based on Audio Signals using Quadratic Discriminant Analysis and High Energy Feature Vectors A. D. Mayvana , S. A. Beheshtib , M. H. Masoomc a Department of Electrical Engineering, Iran University of Science and Technology, Tehran, Iran b Department of Electrical Engineering, Iran University of Science and Technology, Tehran, Iran. c Department of Mechanical Engineering, BabolNoshirvani University of Technology, Babol, Iran. ABSTRACT The focusof this paper is on classification of different vehicles using sound emanated from the vehicles. In this paper,quadratic discriminant analysis classifies audio signals of passing vehicles to bus, car, motor, and truck categories based on features such as short time energy, average zero cross rate, and pitch frequency of periodic segments of signals. Simulation results show that just by considering high energy feature vectors, better classification accuracy can be achieved due to the correspondence of low energy regions with noises of the background. To separate these elements, short time energy and average zero cross rate are used simultaneously.In our method,we have used a few features which are easy to be calculated in time domain and enable practical implementation of efficient classifier. Although, the computation complexity is low, the classification accuracy is comparable with other classification methodsbased on long feature vectors reported in literature for this problem. KEYWORD Classification accuracy; Periodic segments; Quadratic Discriminant Analysis; Separation criterion; Short time analysis. 1. INTRODUCTION Vehicle identification while it is in motion is a prerequisite for traffic and speed management, classified vehicle count, traffic signal time optimization, gap/ headway measurement, and military purposes. Moving vehicles affect the environment in different ways. Vehicle emits heats, sounds, and magnetic field. There are many approaches investigated vehicle identification based on different kinds of signals. Image processing techniques are used to classify vehicles under the real time traffic management and for Intelligent Transportation Systems (ITS). Inductive loops based
  • 2. International Journal on Soft Computing (IJSC) Vol.6, No. 1, February 2015 54 systems are widely used for determination of vehicle counts [1]. The most promising approach for vehicle identification is the one that is based on acoustic signals. Moving vehicles emit characteristic sounds. These sounds are generated from moving parts, frictions, winds, emissions, tires,etc. Assuming that similar vehicles which have the same working conditions generate the same sounds; then these sounds can be used to classify vehicles [2]. Vehicle classification based on sound signals have already attempted by researchers. Paulrajet al. (2013) [3] used autoregressive modeling algorithm for the analysis to extract the features from the recorded vehicle signals. Probabilistic neural network (PNN) models are developed to classify the vehicle type and its distance. Nooralahiyan et al. (1997) [4] used a directional microphone connected to a DAT (Digital Audio Tape) recorder. The digital signal was pre-processed by LPC (Linear Predictive Coding) parameter conversion based on autocorrelation analysis. A Time Delay Neural Network (TDNN) was chosen to classify individual travelling vehicles based ontheir speed independent acoustic signature to four broad categories: buses or Lorries, small or large saloons, various types of motorcycles, and light goods vehicles or vans. Ghiurcau and Rusu(2009) [5] presented a vehicle sound classification system using time encoded signal processing and recognition (TESPAR) method combined with the archetypes technique implemented on the Matlab platform. The experimental results show the efficiency of the TESPAR method when dealing with vehicle sounds. Effectiveness of features set is usually measured by how well they represent the signal; however, features should also satisfy several conditions: their number should be small (less than 15); also,to enable practical implementation they should be easy to calculate. There is a number of feature extraction methods used in vehicle classification. Some of them produce too many features for a single input vector like estimation of Power Spectrum Density [6], [7], [8], and some are too complicated like Principal Component Analysis [9], [7]. Nevertheless, no feature extraction method fulfills all our requirements, and there is no systematic way to select best feature according to our criterion. Harmonic line, Schur coefficients, and MEL filters, were presented in [10] each relating to properties of vehicle audio. Methods were compared in context of their separability and correct classification rate. Aljaafreh and Dong (2010) [2] investigated two feature extraction methods for acoustic signals from moving vehicles. The first one is based on spectrum distribution and the second one on wavelet packet transform. They evaluated the performance of different classifiers such as K-nearest neighbor algorithm (KNN) and support vector machine (SVM). It is found that for vehicle sound data, a discrete spectrum based feature extraction method outperforms wavelet packet transform method. Experimental results verified that support vector machine is an efficient classifier for vehicles.Other spectrum based feature extraction can be found in [11], [12], [13], [14], and [15]. Similarly, Discrete Wavelet Transform (DWT) is used in [16] and [17] to extract features using statistical parameters and energy content of the wavelet coefficients. This paper describes an algorithm to classify audio signals of vehicles. Quadratic discriminant analysis uses feature vectors of periodic segments with elements: short time energy, average zero cross rate, and pitch frequency to distinguish between vehicles’ signals. If we separate feature vectors with high energy from others properly and train classifier with these vectors, better classification accuracy will be achieved due to the correspondence of low energy regions and noises of the background, as simulation results confirm. Obviously high energy feature vectors have less zero cross rate. Considering this point is the basic idea for separation criterion. Our innovation in separation is assuming vectors by energy larger than [], average of all segments’ energy, and ZCR less than [],average of all segments’ ZCR, are high energy elements. We have used three features which are easy to be calculated in time domain, which enable practical implementation of efficient classifier. The cost paid is lower, but still satisfactory,
  • 3. International Journal on Soft Computing (IJSC) Vol.6, No. 1, February 2015 55 classification accuracy than other classification applications reported in literature for this problem. Sections of paper are as follow: After introduction first,we determine important parameters and features.Then we study quadratic discriminant analysis and high energy element separation criterion in section III. Simulation results show improvement in classification accuracy for separation criterion. Finally, the paper ends by conclusion. 2. FEATURE EXTRACTION Changing the characteristics of an audio signal over time seems normal. For example, Changes in the signal’s peak value or rate of sign changes of amplitude are such a time varying features.Although these changes exist in signals, short time analysis is a method to illustrate them asobvious way. Basic assumption of this method is based on these changes occur slowly; then wedivide signal into smaller parts and features are extracted. 2.1. Short Time Energy This feature indicates how the signal amplitude changes over the time which is defined as follows: = ∑ [
  • 4. − ] ∞ ∞ (1) Where is the audio signal and
  • 5. is the window that slides along the audio sequence selecting the interval to be involved in the computation. 2.2. Short Time Average Zero Cross Rate This feature indicates how the signal sign changes over the time which is defined as follows: = ∑ | [ ] − [ − 1]| ∞ ∞
  • 7. = 1 2 ⁄ 0 ≤ ≤ − 1 = 0 ℎ !
  • 8. (3) The rate at which signal sign changes occur is a simple measure of the frequency content of a signal. This is particularly true for narrowband signals. For example, a sinusoidal signal of frequency #$, sampled at a rate#%, has 2 #$ #% ⁄ average rate of zero-crossings.Thus, the average ZCR gives a resonable way to estimate the frequency of a sine wave. Audio signals are broadband signals and the interpretation of average ZCR is therefore much less precise. However, rough estimates of spectral properties can be obtained using a represantation based onthe short time energy and zero cross rate together.This representation was proposed by Reddy, and studied by Vicens [18] as the basis for a large-scale speech recognition system.
  • 9. International Journal on Soft Computing (IJSC) Vol.6, No. 1, February 2015 56 2.3. Pitch Frequency Pitch period estimation (or fundamental frequency) is one of the most important problems in speech processing. Pitch detectors are used in vocoders [19], speaker identification and verification systems [20, 21]. Because of its importance, many solutions to this problem have been proposed [22]. In this paper we use the fact thatthe autocorrelation function for periodic segments attains a maximum at samples 0, ±',±2',… where T is pitch period. Since, there may be exist other autocorrelation peaks except those due to the periodicity, the simple procedure of picking the largest peak at sample'in the autocorrelation will failto estimate Pitch period. To avoid this problem it is useful to process the signal so as to make periodicity more prominent while suppressing other distracting features of the signal. Spectrum flattener with the objective to bring each harmonic to the same amplitude level as in the case of a periodic impulse trainhas been used. Specifically, center clipped signalhas been used in computing autocorrelation function. In the scheme proposed,similar to [23], the center clipped signal is obtained by following nonlinear transformation For samples above (), the output is equal to the input minus the clipping level. For samples below the clipping level the output is zero.We set clipping level, (), equal to 68% of the minimum of two maximum amplitudes found in both the first third and last third of the audiosegment;denoted respectively by *+ and *. However, for the un-periodic segment, there are no strong autocorrelation periodicity peaks and it seems to be a high frequency noise-like waveform [22]. We assume segments with weak peaks below 30 % clipped signal’s energy, ,)-. = ∑ [,)-.
  • 10. − ] ∞ ∞ ,are un-periodic. Finally pitch frequencies of frames smoothed through median operation. Following figures present features. −()() Fig.1: center clipping function
  • 11. International Journal on Soft Computing (IJSC) Vol.6, No. 1, February 2015 57 Fig.2: Energy for Bus in first row for three differentsounds, for Car in the second row, for Motor in the third row, and for Truck in the fourth row. Fig.3:ZCR for Bus in first row for three different sounds, for Car in the second row, for Motor in the third row, and for Truck in the fourth row.
  • 12. International Journal on Soft Computing (IJSC) Vol.6, No. 1, February 2015 58 Fig.4: Pitch Frequency for Bus in first row for three different sounds, for Car in the second row, for Motor in the third row, and for Truck in the fourth row. Features are extracted from rectangular windowed segment of length 15 Milliseconds and overlapping of 5 Milliseconds between consecutive windows. Feature vectors of periodic segments are used for training and classification. After determining the important features and parameters for audio signal modeling, we will review quadratic discriminant analysis and high energy separation criterion will be described. 3. CLASSIFIER AND SEPARATION CRITERION Now the discriminant boundaries between classes should be determined. Quadratic discriminantanalysis is used to discriminate the classes. The classifier that separate K classes has K quadratic functions as follow [24]: /0 1 = 12 31 + 52 1 + 6$ (4) The coefficient6$is constant coefficient,5is vector of linear coefficients, and 3 is matrix containing quadratic coefficients. Feature vectorX will belong to class k, if/0 1 /8 1; ∀; ≠ =.In classification step, after feature vectors were extracted, we are able to define class of each vector separately. Finally, signal will allocate to a classwhichachieved great supply in feature vector classification phase. In other words, probability of belonging to each class is proportional to abundance of that class in allocated labels to vectors by quadratic discriminant analysis.
  • 13. International Journal on Soft Computing (IJSC) Vol.6, No. 1, February 2015 59 3.1. High Energy Vector Separation Criterion Up to now all periodic vectors are considered in training and classification. However, low energy regions corresponding with noises of the background consist of high frequency variations may not be suited for vehicle classification. Hence by ignoring mentioned vectors, an improvement in classification accuracy can be achieved.On the other hand, it can be observed that when amplitude is high zero cross rate decreases. So, high energy elements have low zero cross rate, and low energy vectors have fast sign alternation. A criterion for separation elements with high energy and low zero cross rate can be defined as follow: . []And A. [] (5) Where []and []denote average of all segments’ energy, ZCR. The coefficientα is more than 1, andAis less than 1. The more α and the less A than unit, the less data coincides in condition. In this paper α = A = 1are chosen. Figure 5 presents a block diagram of our method. Fig.5: Block diagram of the proposed scheme for vehicle classification
  • 14. International Journal on Soft Computing (IJSC) Vol.6, No. 1, February 2015 60 4. RESULTS AND DISCUSSION Simultaneous arrival of different types vehicles, overtaking at the study location, and random back ground noises were the different challenges encountered in the data collection. Sound emitted by vehicles was captured for a divided road carrying moderate traffic. Roadside recording was performed on clear sunny working days. The recordings’ sites were selected to capture the sound of moving vehicles without much acceleration change.Environmental back ground noise is at moderate level. The conditions under which data set is recorded are realistic.Our algorithm is implemented on the Matlab7.10.0 (R2010a) platform. Preprocessing: Each signal is normalized such that −1 ≤ ≤ 1. Windowed signal segments are filtered by 4DE degree Butterworth low pass filter. The data set includes four classes; bus, car, motor, and truck. All signals sampled at 11025 rate. Each signal is segmented by rectangular window of length 15 milliseconds and overlapping of 5 milliseconds between consecutive windows. Vectors containing short time energy, average zero cross rate, and pitch frequency are calculated for each segment and those with 0 pitch frequencies are removed. We used k-Fold cross validation with k=10 in our applications. In this method data is randomly divided into a test and training setk different times [25] and the average error across allk trials is computed. Table 1 provides the values of some important parameters employed in this work. Classification accuracies for following classifier have been provided in table 2. • least square classifier, • k-nearestneighbor classifier, • quadratic and linear discriminant analysiswhen training and testing have been done just based on vectors which be able to satisfy the (5) conditions;denoted respectively by 3FG∗∗ and IFG∗∗ , • SVM classifier with harmonic line based feature vectors [10], • SVM classifier with Schur filter coefficients based feature vectors [10], • SVM classifier with Mel filter coefficients based feature vectors [10]. We see improvement in classification accuracy has been achieved forquadratic discriminant analysis and separation criterion, 3FG∗∗ . The correct classification rate depends not only on the feature extraction method, but also on the type of the classifier. This is the reason that why not we see a considerable improvement for IFG∗∗ . Table 3 shows computation complexity in terms of feature space dimensionality for our proposed algorithm and last 3 methods in table 2. In comparison, we should note that the accuracy of 3FG∗∗ with 3 short time parameters based feature vectors is comparable with SVM with 16 Schur coefficients. SVM with 12 Mel filter coefficients accuracy is 80% and accuracy of QDA∗∗ is 68.75%. Although the computation complexity of QDA∗∗ is low, but the classification accuracy is lower thanSVM with Mel filter coefficients.
  • 15. International Journal on Soft Computing (IJSC) Vol.6, No. 1, February 2015 61 Signal sample rate 11025 sample/sec Length of window for classification 165 sample Overlap between windows for classification 55 sample Type of window Rectangular Preprocessor 4DE degree ButterworthLPF LPF’s cut off frequency 4 KHz Clipping level 0.68 ∗ min { *+, *} Maximum pitch frequency f= 1000 Hz Periodic/Un-periodic threshold 0.3 ∗ ,)-. Number of frames over which median operates 3 α and ζ 1 Number of cross fold validation 10 examples per class: [bus, car, motor, truck] [50, 70, 80, 40] Table 1: Values of some important parameters Classifier Accuracy (%) Least Square 30 kNN, k=25, Cosine 52.09 ± 2.2 kNN, k=25, Euclidian 53.33 ± 3.83 LDA 46.66 IFG∗∗ 49.58 QDA 56.25 3FG∗∗ 68.75 SVM with harmonic line 83 SVM with Schur coefficients 68 SVM with Mel filter coefficients 80 Table 2: Classification accuracy
  • 16. International Journal on Soft Computing (IJSC) Vol.6, No. 1, February 2015 62 Feature extraction method Number of features Short time analysis 3 harmonic lines 5 Schur coefficients 16 Mel filter coefficients 12 Table 3: Number of features Additionally confusion matrix, contain information about actual and predicted classifications done by 3FG∗∗ classification system, shown in table 4. Actual Predicted Bus vehicles Car vehicles Motor vehicles Truck vehicles Bus vehicles 27 20 3 0 Car vehicles 0 70 0 0 Motor vehicles 16 12 48 4 Truck vehicles 9 11 0 20 Table 4: Representation of confusion matrix Although recording conditions were realistic, but we should examine proposed algorithm when environmental noises increase like rainy days.Since we use high energy feature vectors, 3FG∗∗ can save its effectiveness. CONCLUSION In the research reported in this paper, QDA with a criterion to separate high energy feature vectors was applied on the task of classification of vehicles based on audio signals and the simplest method with satisfactory accuracy was evaluated. Short time energy, average zero cross rate, and pitch frequency made feature vectors; then,signals were classified using quadratic discriminant analysis. We found that classification accuracy improved from 56.25% to 68.75%just by considering components which have (5) conditionsdue to the correspondence of low energy regions and noises of the background.At last, the obtained accuracy was 68.75% by using three features while the 80% correct classification rate was achievedby SVM with 12Mel coefficients [10].The results strongly suggest that proposed method can aid the practical implementation.
  • 17. International Journal on Soft Computing (IJSC) Vol.6, No. 1, February 2015 63 REFERENCES [1] J. George, and et al. ‘’ Exploring Sound Signature for Vehicle Detection and ClassificationUsing ANN’’ International Journal on Soft Computing (IJSC) Vol.4, No.2, May 2013. [2] A. Aljaafreh, and L. Dong ‘’an Evaluation of Feature Extraction Methods for VehicleClassification Based on Acoustic Signals’’ International Conference on Networking, Sensing andControl (ICNSC), 2010. [3] M.P. Paulraj, and et al. ‘’Moving Vehicle Recognition and Classification Based on TimeDomain Approach‘’ Procedia Engineering, Volume 53, 2013, Pages 405–410. [4] Y. Nooralahiyan, and et al. ‘’Field Trial of Acoustic Signature Analysis for VehicleClassification’’ Transportation Research Part C: Emerging Technologies, Volume 5, Issues 3–4,August–October 1997, Pages 165–177. [5] M. V. Ghiurcau, C. Rusu, ‘’Vehicle Sound Classification Application and Low PassFiltering Influence’’ In proceeding of International Symposium on Signals, Circuits andSystems, 2009, ISSCS 2009. [6] M. Wellman, N. Srour, and D. Hillis, “Feature Extraction and Fusion of Acoustic andSeismic Sensors for Target Identification,” in Society of Photo-Optical InstrumentationEngineers (SPIE) Conference Series, ser. Society of Photo-Optical Instrumentation Engineers(SPIE) Conference Series, G. Yonas, Ed, vol. 3081. SPIE, 1997, pp. 139–145. [7] M. Wellman, “Acoustic Feature Extraction for a Neural Network Classifier.” DTICDocument, Tech. Rep., 1997. [8] M. F. Duarte and Y. H. Hu, “Vehicle Classification in Distributed Sensor Networks,” Journalof Parallel and Distributed Computing, vol. 64, pp.826–838, 2004. [9] H. Wu, M. Siegel, and P. Khosla, “Vehicle Sound Signature Recognition by FrequencyVector Principal Component Analysis,” in Instrumentation and Measurement TechnologyConference, 1998. IMTC/98.Conference Proceedings.IEEE, vol. 1, May 1998, pp. 429 –434vol.1. [10] M. Gorski, J. Zarzycki, ‘’Feature Extraction in Vehicle Classification’’ InternationalConference on Signals and Electronic Systems (ICSES), 2012. [11] H. Wu, M. Siegel, and P. Khosla, “Distributed classification of acoustictargets wireless audio-sensor networks,” Computer Networks, vol. 52,no. 13, pp. 2582–2593, Sep. 2008. [12] ——, “Vehicle classification in distributed sensor networks,” Journalof Parallel and Distributed Computing, vol. 64, no. 7, pp. 826–838,July 2004. [13] Y. Seung S., K. Yoon G., and H. Choi, “Distributed and efficientclassifiers for wireless audio-sensor networks,” in 5th InternationalConference on Volume, Apr. 2008. [14] S. S. Yang, Y. G. Kim1, and H. Choi, “Vehicle identification usingdiscrete spectrums in wireless sensor networks,” Journal of Networks,vol. 3, no. 4, pp. 51–63, Apr. 2008. [15] H. Wu, M. Siegel, and P. Khosla, “Vehicle sound signature recognitionby frequency vector principal component analysis,” IEEE Trans.Instrum. Meas., vol. 48, no. 5, pp. 1005–1009, Oct. 1999. [16] C. H. C. K. R. E. G. G. R. and M. T. J, “Wavelet-based ground vehiclerecognition using acoustic signals,” Journal of Parallel and DistributedComputing, vol. 2762, no. 434, pp. 434–445, 1996. [17] A. H. Khandoker, D. T. H. Lai, R. K. Begg, and M. Palaniswami,“Wavelet-based feature extraction for support vector machines forscreening balance impairments in the elderly,” vol. 15, no. 4, pp. 587– 597, 2007. [18] P. J. Vicens, “Aspects of Speech Recognition by Computer,” Ph.D. Thesis, Stanford Univ., AI Memo No. 85, Comp. Sci. Dept., Stanford Univ., 1969. [19] J. L. Flanagan, Speech Analysis, Synthesis and Perception, 2nd Ed., Springer Verlag, N.Y., 1972. [20] B. S. Atal, “Automatic Speaker Recognition Based on Pitch Contours,” J. Acoust. Soc. Am., Vol. 52, pp. 1687-1697, December 1972. [21] A. E. Rosenberg and M. R. Sambur, “New Techniques for Automatic Speaker Verification,” IEEE Trans. Acoust, Speech, and Signal Proc., Vol.23, pp. 169-176, April 1975. [22] L. R. Rabiner, R. W. Schafer, Digital Processing of Speech Signals. Englewood Cliffs, N.J., Prentice Hall.
  • 18. International Journal on Soft Computing (IJSC) Vol.6, No. 1, February 2015 64 [23] M. M. Sondhi, “New Methods of Pitch Extraction,” IEEE Trans. Audio and Electroacoustics, Vol. 16, No. 2, pp. 262-266, June 1968. [24] K. Fukunaga, Introduction to Statistical Pattern Recognition. San Diego, Academic Press,1990, pp. 153-154. [25] R. Kohavi, F. Provost, Glossary of terms, Editorial for the Special Issue on Applications ofMachine Learning and the Knowledge Discovery Process, vol. 30, No. 2–3, 1998.