Music Genre Classification Using Ga-Induced Essay

Music genre is frequently used as the query of choice while browsing through the music databases over the Internet[l]. While research onto acquiring suitable features and classifiers for genre classification is [4], [5]), works involving selection of relevant feature-sets particular to the task at hand are sparse at best. Some examples of such work would be [6], [7], where techniques of PICA, OFFS and BBS have been used effectively. In our work, we investigate the possibility of a relevant and minimal feature-set-selection with the help of Genetic Algorithms (GA).

GAS have been used before by [8] and [2] to efficiently classify music genres. While [2] has shown pretty impressive results (a 90% classification rate) using GA, 09 musical features and a hierarchical classifier, [8] has been able to get almost 60% classification rate with GA, a set of classifiers like ann., SAVE and a feature set of 30 features. These works, though impressive, aim only for an optimal feature-set devoid of noisy features(which tend to misclassifying samples). Furthermore, they use a primitive GA fitness function, biz. He hit- rate, for the best feature-set selection, which might converge to a local minima instead of a global one In this work, we not only find the optimal features from a set of given features for a given classification problem, but we also minimize the number of features required for the task. Using a modified fitness function, that also eliminates the local-minima problem of the standard hit-rate one, we bring about a more than 80% reduction in the original feature set, and a 50% reduction in the feature-set derived from the original feature-set by the hit-rate-only fitness function.

To reiterate, we haven’t proposed a new method for classification, or a new method for feature generation. We assume that the feature set and the classifier are defined by the task at hand. Given this scenario, and the fact hat the standard GA, or other feature reduction technique uses have already been employed to reduce the original feature-set, can we further reduce the number of features with minimal loss in performance? – thigh’s the question we are going to handle in this paper. The paper is organized as follows.

In Section II, we give a brief description of the established GA and feature selection methods in literature, and propose our modification. Then in Section Ill, we describe the dataset, the feature-set and classifiers we are using for the experiment. Then in Section IV, we discuss the findings of our test and validate our claims. Then we conclude our discussion in Section V. II. P REPOSED M DEL Given a set of possible features and a classifier for a classification task, our objective is to derive an optimal features devoid of noisy features with an aim of minimizing its cardinality.

Using GAS to derive near-optimal feature-subsets was brought to light by [1 0], and was further investigated in [1 1], [9]. In the present work, extending [10] and we propose a modified GA to derive a near-optimal-UCM-minimal featureless for music genre classification. GAS emulate human evolutionary characteristics, and follow the survival-of-the-fittest paradigm. In an personalized model of GA, we have a population of chromosomes (essentially a bit string), which are assigned fitness values (their relative worth) based on the task at hand.

Fit individuals/chromosomes mate with each other, and that is simulated by crossover, in which they exchange parts of their chromosome. To account for sudden characteristic changes in some individuals of a population, mutation is simulated through bit flipping random parts of the chromosome. The fit individuals are selected from a population using some selection rule and the next generation consists of these fit individuals whereas unfit ones are discarded.

With this approach, the GA claims that at the end of hundreds of generations, only the best chromosomes for the task at hand will remain, and that is the assumption based on which best features are selected. The crossover rate, mutation rate, selection rule, fitness function etc. Are determined by the problem at hand. For an n-dimensional feature space, we choose n-bit chromosomes. Following [1 0], a bit-value of ‘1’ means that the feature-dimension is considered for classification, and a value of ‘O’ means that the feature is eliminated from consideration.

The traditional fitness-function is the hit-rate (HER), I. E. The fraction of train-set samples that were correctly classified using the selected features. But such a fitness function has a tendency to converge to a local minima, and it also disregards the objective of creating the maximum separation between the classes under consideration. So, following we use the following fitness function: F tiniest = HER + y(CD/p) where CD is the class distance, p is the total number of samples, and y is a scaling constant. The class distance is defined as follows, p p Ill.

0 Comment