Skip to Main content Skip to Navigation
Conference papers

AdaBoost Parallelization on PC Clusters with Virtual Shared Memory for Fast Feature Selection

Abstract : Feature selection is a key issue in many machine learning applications and the need to test lots of candidate features is real while computational time required to do so is often huge. In this paper, we introduce a parallel version of the well- known AdaBoost algorithm to speed up and size up feature selection for binary classification tasks using large training datasets and a wide range of elementary features. This parallelization is done without any modification to the AdaBoost algorithm and designed for PC clusters using Java and the JavaSpace distributed framework. JavaSpace is a memory sharing paradigm implemented on top of a virtual shared memory, that appears both efficient and easy-to-use. Results and performances on a face detection system trained with the proposed parallel AdaBoost are presented.
Complete list of metadata

Cited literature [8 references]  Display  Hide  Download
Contributor : Sébastien van Luchene Connect in order to contact the contributor
Submitted on : Friday, January 25, 2008 - 9:48:59 PM
Last modification on : Monday, December 6, 2021 - 6:08:02 PM
Long-term archiving on: : Thursday, April 15, 2010 - 4:07:25 PM


Publisher files allowed on an open archive


  • HAL Id : hal-00216041, version 1



Virginie Galtier, Olivier Pietquin, Stéphane Vialle. AdaBoost Parallelization on PC Clusters with Virtual Shared Memory for Fast Feature Selection. IEEE International Conference on Signal Processing and Communication, Nov 2007, Dubai, United Arab Emirates. pp.165-168. ⟨hal-00216041⟩



Les métriques sont temporairement indisponibles