AdaBoost Parallelization on PC Clusters with Virtual Shared Memory for Fast Feature Selection

Abstract : Feature selection is a key issue in many machine learning applications and the need to test lots of candidate features is real while computational time required to do so is often huge. In this paper, we introduce a parallel version of the well- known AdaBoost algorithm to speed up and size up feature selection for binary classification tasks using large training datasets and a wide range of elementary features. This parallelization is done without any modification to the AdaBoost algorithm and designed for PC clusters using Java and the JavaSpace distributed framework. JavaSpace is a memory sharing paradigm implemented on top of a virtual shared memory, that appears both efficient and easy-to-use. Results and performances on a face detection system trained with the proposed parallel AdaBoost are presented.
Complete list of metadatas

Cited literature [8 references]  Display  Hide  Download

https://hal-supelec.archives-ouvertes.fr/hal-00216041
Contributor : Sébastien van Luchene <>
Submitted on : Friday, January 25, 2008 - 9:48:59 PM
Last modification on : Thursday, March 29, 2018 - 11:06:04 AM
Long-term archiving on : Thursday, April 15, 2010 - 4:07:25 PM

File

Supelec334.pdf
Publisher files allowed on an open archive

Identifiers

  • HAL Id : hal-00216041, version 1

Collections

Citation

Virginie Galtier, Olivier Pietquin, Stéphane Vialle. AdaBoost Parallelization on PC Clusters with Virtual Shared Memory for Fast Feature Selection. IEEE International Conference on Signal Processing and Communication, Nov 2007, Dubai, United Arab Emirates. pp.165-168. ⟨hal-00216041⟩

Share

Metrics

Record views

331

Files downloads

326