Feature selection based on information theory

Empreu sempre aquest identificador per citar o enllaçar aquest ítem http://hdl.handle.net/10045/18362
Registre complet
Registre complet
Camp Dublin Core Valor Idioma
dc.contributor.advisorCazorla Quevedo, Miguel Ángel-
dc.contributor.advisorEscolano Ruiz, Francisco-
dc.contributor.authorBonev, Boyan-
dc.contributor.otherUniversidad de Alicante. Departamento de Ciencia de la Computación e Inteligencia Artificialen
dc.date.accessioned2011-07-27T12:02:52Z-
dc.date.available2011-07-27T12:02:52Z-
dc.date.created2010-
dc.date.issued2010-
dc.date.submitted2010-06-29-
dc.identifier.isbn978-84-694-1634-1-
dc.identifier.urihttp://hdl.handle.net/10045/18362-
dc.description.abstractAlong with the improvement of data acquisition techniques and the increasing computational capacity of computers, the dimensionality of the data grows higher. Pattern recognition methods have to deal with samples consisting of thousands of features and the reduction of their dimensionality becomes crucial to make them tractable. Feature selection is a technique for removing the irrelevant and noisy features and selecting a subset of features which describe better the samples and produce a better classification performance. It is becoming an essential part of most pattern recognition applications.en
dc.description.abstractIn this thesis we propose a feature selection method for supervised classification. The main contribution is the efficient use of information theory, which provides a solid theoretical framework for measuring the relation between the classes and the features. Mutual information is considered to be the best measure for such purpose. Traditionally it has been measured for ranking single features without taking into account the entire set of selected features. This is due to the computational complexity involved in estimating the mutual information. However, in most data sets the features are not independent and their combination provides much more information about the class, than the sum of their individual prediction power.en
dc.description.abstractMethods based on density estimation can only be used for data sets with a very high number of samples and low number of features. Due to the curse of dimensionality, in a multi-dimensional feature space the amount of samples required for a reliable density estimation is very high. For this reason we analyse the use of different estimation methods which bypass the density estimation and estimate entropy directly from the set of samples. These methods allow us to efficiently evaluate sets of thousands of features.en
dc.description.abstractFor high-dimensional feature sets another problem is the search order of the feature space. All non-prohibitive computational cost algorithms search for a sub-optimal feature set. Greedy algorithms are the fastest and are the ones which incur less overfitting. We show that from the information theoretical perspective, a greedy backward selection algorithm conserves the amount of mutual information, even though the feature set is not the minimal one.en
dc.description.abstractWe also validate our method in several real-world applications. We apply feature selection to omnidirectional image classification through a novel approach. It is appearance-based and we select features from a bank of filters applied to different parts of the image. The context of the task is place recognition for mobile robotics. Another set of experiments are performed on microarrays from gene expression databases. The classification problem aims to predict the disease of a new patient. We present a comparison of the classification performance and the algorithms we present showed to outperform the existing ones. Finally, we succesfully apply feature selection to spectral graph classification. All the features we use are for unattributed graphs, which constitutes a contribution to the field. We also draw interesting conclusions about which spectral features matter most, under different experimental conditions. In the context of graph classification we also show important is the precise estimation of mutual information and we analyse its impact on the final classification results.en
dc.languageengen
dc.publisherUniversidad de Alicanteen
dc.subjectFeature selectionen
dc.subjectInformation theoryen
dc.subjectPattern recognitionen
dc.subjectSupervised classificationen
dc.subjectSelección de característicasen
dc.subjectTeoría de la informaciónen
dc.subjectReconocimiento de patronesen
dc.subjectClasificación supervisadaen
dc.subject.otherCiencia de la Computación e Inteligencia Artificialen
dc.titleFeature selection based on information theoryen
dc.typeinfo:eu-repo/semantics/doctoralThesisen
dc.rights.accessRightsinfo:eu-repo/semantics/openAccess-
Apareix a la col·lecció: Tesis doctorals

Arxius per aquest ítem:
Arxius per aquest ítem:
Arxiu Descripció Tamany Format  
Thumbnailtesis_Ivanov.pdf3,63 MBAdobe PDFObrir Vista prèvia


Aquest ítem està subjecte a una llicència de Creative Commons Llicència Creative Commons Creative Commons