Show simple item record

Combining Variable Selection with Dimensionality Reduction

dc.date.accessioned2005-12-22T02:25:27Z
dc.date.accessioned2018-11-24T10:24:26Z
dc.date.available2005-12-22T02:25:27Z
dc.date.available2018-11-24T10:24:26Z
dc.date.issued2005-03-30
dc.identifier.urihttp://hdl.handle.net/1721.1/30531
dc.identifier.urihttp://repository.aust.edu.ng/xmlui/handle/1721.1/30531
dc.description.abstractThis paper bridges the gap between variable selection methods (e.g., Pearson coefficients, KS test) and dimensionality reductionalgorithms (e.g., PCA, LDA). Variable selection algorithms encounter difficulties dealing with highly correlated data,since many features are similar in quality. Dimensionality reduction algorithms tend to combine all variables and cannotselect a subset of significant variables.Our approach combines both methodologies by applying variable selection followed by dimensionality reduction. Thiscombination makes sense only when using the same utility function in both stages, which we do. The resulting algorithmbenefits from complex features as variable selection algorithms do, and at the same time enjoys the benefits of dimensionalityreduction.1
dc.format.extent10 p.
dc.format.extent14957523 bytes
dc.format.extent722450 bytes
dc.language.isoen_US
dc.subjectAI
dc.subjectComputer Vision
dc.subjectStatistical Learning
dc.subjectVariable Selection
dc.titleCombining Variable Selection with Dimensionality Reduction


Files in this item

FilesSizeFormatView
MIT-CSAIL-TR-2005-019.pdf722.4Kbapplication/pdfView/Open
MIT-CSAIL-TR-2005-019.ps14.95Mbapplication/postscriptView/Open

This item appears in the following Collection(s)

Show simple item record