• Ei tuloksia

In this last section of the pre-processing stage, we generated a multitude of feature sub-sets which became the inputs to the classification stage. We created feature subsets in three ways, manually, semi-manually and algorithmically. In this section, we elaborate on each feature selection strategy. The entire feature selection and sub-set generation pipeline is shown in figure 15 at the end of the section.

3.4.1 Manual Selection

In the manual selection case, we devised two feature sets and further generated more sub-sets between statistical summaries; When selecting feature mean values we code-name the sub-set as โ€˜Feature sub-set (๐œ‡)โ€™, when we select the standard deviation values we used the name โ€˜Feature sub-set (๐œŽ)โ€™, when both summaries were used, we aggregated both code-names as โ€˜Feature sub-set (๐œ‡, ๐œŽ)โ€™. In tables 15 and 16 we show each manually selected feature sub-set, dimensionality and description.

TABLE 15. The โ€˜All Featuresโ€™ sets, comprising of all features and every summary statistic sub-set.

โ€˜All Featuresโ€™ Sets Dimensionality Description

โ€˜All Features (๐)โ€™ 63 Only mean values of all features.

โ€˜All Features (๐ˆ)โ€™ 63 Only standard deviation values of all features.

โ€˜All Features (๐, ๐ˆ)โ€™ 126 Mean and standard deviation values of all features.

TABLE 16. The โ€˜individual featureโ€™ sets, comprising of each individual feature and both summary statistics.

โ€˜Individual featureโ€™ Sets Dimensionality Description

โ€˜SB-Entropy (๐, ๐ˆ)โ€™ 20 Mean and standard deviation values of 10 sub-band entropies.

โ€˜SB-Flux (๐, ๐ˆ)โ€™ 20 Mean and standard deviation values of 10 sub-band

fluxes.

โ€˜SB-Kurtosis (๐, ๐ˆ)โ€™ 20 Mean and standard deviation values of 10 sub-band kurtosis.

โ€˜SB-ZCR (๐, ๐ˆ)โ€™ 20 Mean and standard deviation values of 10 Sub-Band

zero crossing rates.

โ€˜SB-Skewness (๐, ๐ˆ)โ€™ 20 Mean and standard deviation values of 10 sub-band skewness.

โ€˜MFCCs (๐, ๐ˆ)โ€™ 26 Mean and standard deviation values of 13 MFCC

coefficients.

3.4.2 Semi - Manual Selection

In semi-manual selection, we devise the โ€˜Top 2โ€™ feature sub-set from the classification performance ranking of the โ€˜individual featuresโ€™ sets. This selection design deals with two features only, for this reason, the ๐œ‡1, ๐œŽ1 codes referred to the mean and standard deviation of the best performing individual feature, while ๐œ‡2, ๐œŽ2 referred to the same statistics for the second-best feature. Table 17 displays the composition of our โ€˜Top 2โ€™ semi- manual selection design.

TABLE 17. The semi-manual feature selection subsets, containing the top two performing individual features.

Top 2 Feature Sub-Set

Algorithmic feature selection or simply โ€˜feature selectionโ€™ plays an important role in automatically ranking and selecting relevant features for a classification task. Often a portion of features involved in machine learning may not be informative or relevant to the classification task. Indeed, a high number of irrelevant features may increase the complexity of a model and even decrease classification and computational performance by increasing overfitting or imposing other unwanted effects. Feature selection algorithms help to combat such effects. Importantly, feature selection should not be perplexed with dimensionality reduction methods, both reduce the number of features for a given task, but dimensionality reduction methods may produce different features from the initial feature set, feature selection methods do not.

There are three types of feature selection methods, filter methods, wrapper methods and embedded methods. Filter methods employ statistical measures to score each feature with

respect to the dependent variable or independently. In contrast, wrapper methods generate feature combinations and compare the classification accuracy results of such combinations directly via the classification stage. Embedded methods identify feature contributions to classification accuracy within and during the classification process. A popular class of embedded methods is regularization techniques often used to penalize classification and regression algorithms such that they may reduce over-reliance on specific features, this often may lead to reduced overfitting.

The choice of feature selection algorithm depends heavily on the understanding of a given problem. There is no โ€˜one size fits allโ€™ selection algorithm, often selection algorithms tend to produce completely different rankings for the same input features. In the absence of deep problem understanding it is common for multiple selection methods to be evaluated, and at times aggregated ranks between multiple rankings may be performed. In our study, we employ a filter method using the Information gain (IG) algorithm that we detailed below. The resulting feature sub-set was code-named โ€˜Information Gain Top 20โ€™ to refer to the top 20 features suggested by IG when inputting the โ€˜All Features (๐œ‡, ๐œŽ)โ€™ feature set. We implemented IG in python with the Orange3 library (Demลกsar et al., 2013).

3.4.4 Information Gain

Information Gain (IG) is a filter method computed for each feature with respect to the class labels (Liu & Motoda, 1998); it relies heavily on the Shannonโ€™s information entropy (Shannon, 2001). To obtain ๐ผ๐บ(๐‘‹|๐‘Œ) were ๐‘‹ and ๐‘Œ are random variables, let us consider the formula for the information entropy ๐ป of variable ๐‘‹:

๐ป(๐‘‹) = โˆ’ โˆ‘ ๐‘(๐‘ฅ)

๐‘ฅโˆˆ๐‘‹

๐‘™๐‘œ๐‘”2(๐‘(๐‘ฅ))

Where ๐‘(๐‘ฅ) is the marginal probability density function for ๐‘‹. When we introduce variable ๐‘Œ and devise the values of variable ๐‘‹ with respect to the values of ๐‘Œ, a relationship between ๐‘‹ and ๐‘Œ exists only when the entropy of ๐‘‹ devised by ๐‘Œ is smaller than the initial entropy of ๐‘‹, the entropy of ๐‘‹ devised by ๐‘Œ is given by:

๐ป(๐‘‹|๐‘Œ) = โˆ’ โˆ‘ ๐‘(๐‘ฆ)

๐‘ฆโˆˆ๐‘Œ

โˆ‘ ๐‘(๐‘ฅ|๐‘ฆ)๐‘™๐‘œ๐‘”2(๐‘(๐‘ฅ|๐‘ฆ))

๐‘ฅโˆˆ๐‘‹

Where ๐‘(๐‘ฆ) is the marginal probability density function for variable ๐‘Œ and ๐‘(๐‘ฅ|๐‘ฆ) is the conditional probability of ๐‘ฅ given ๐‘ฆ. After the conditional entropy step, information gain is defined as the amount of entropy decrease in ๐‘‹ represented by the surplus information provided from ๐‘Œ about ๐‘‹, formally defined as:

๐ผ๐บ(๐‘‹|๐‘Œ) = ๐ป(๐‘‹) โˆ’ ๐ป(๐‘‹|๐‘Œ)

Importantly, ๐ผ๐บ(๐‘‹|๐‘Œ) = ๐ผ๐บ(๐‘Œ|๐‘‹) because information gain is symmetrical for the two variables (Yu & Liu, 2003).

3.4.5 Feature Selection Overview

To summarize, figure 15 maps the flow of our feature selection sets, along with their statistical summary subsets with respect to the classification stage.

FIGURE 15. The flow of feature selection subsets with respect to the classification stage.