Abstract
Genetic algorithms are powerful tools for k-nearest neighbors classification. Traditional knn classifiers employ Euclidian distance to assess neighbor similarity, though other measures may also be used. GAs can search for optimal linear weights of features to improve knn performance using both Euclidian distance and cosine similarity. GAs also optimize additive feature offsets in search of an optimal point of reference for assessing angular similarity using the cosine measure. This poster explores weight and offset optimization for knn with varying similarity measures, including Euclidian distance (weights only), cosine similarity, and Pearson correlation. The use of offset optimization here represents a novel technique for enhancing Pearson/knn classification performance. Experiments compare optimized and non-optimized classifiers using public domain datasets. While unoptimized Euclidian knn often outperforms its cosine and Pearson counterparts, optimized Pearson and cosine knn classifiers show equal or improved accuracy compared to weight-optimized Euclidian knn.
Original language | English |
---|---|
Title of host publication | GECCO 2005 - Genetic and Evolutionary Computation Conference |
Editors | Hans-Georg Beyer |
Place of Publication | New York, NY |
Publisher | Association for Computing Machinery |
Pages | 1549-1550 |
Number of pages | 2 |
ISBN (Print) | 978-1-59593-010-1 |
DOIs | |
State | Published - 2005 |
Event | GECCO 2005 - Genetic and Evolutionary Computation Conference - Washington, D.C., United States Duration: Jun 25 2005 → Jun 29 2005 |
Conference
Conference | GECCO 2005 - Genetic and Evolutionary Computation Conference |
---|---|
Country/Territory | United States |
City | Washington, D.C. |
Period | 6/25/05 → 6/29/05 |
ASJC Scopus Subject Areas
- General Engineering
Keywords
- Dimensionality reduction
- Genetic algorithms
- K-nearest neighbors
- Pattern recognition
Disciplines
- Computational Biology