TY - GEN
T1 - Distributed mixture-of-experts for Big Data using PETUUM framework
AU - Peralta, Billy
AU - Parra, Luis
AU - Herrera, Oriel
AU - Caro, Luis
N1 - Funding Information:
This work was partially financed by FONDECYT Initiation project 11140892.
PY - 2018/7/5
Y1 - 2018/7/5
N2 - Today, organizations are beginning to realize the importance of using as much data as possible for decision-making in their strategy. The finding of relevant patterns in enormous amount of data requires automatic machine learning algorithms, among them, a popular option is the mixture-of-experts that allows to model data using a set of local experts. The problem of using typical learning algorithms over Big Data is the handling of these large datasets in primary memory. In this paper, we propose a methodology to learn a mixture-of-experts in a distributed way using PETUUM platform. Particularly, we propose to learn the parameters of mixture-of-experts by adapting the standard stochastic gradient descent in a distributed way. This methodology is applied to people detection with standard real datasets considering accuracy and precision metrics among other. The results show a consistent performance of mixture-of-experts models where the best number of experts varies according to the particular dataset. We also evidence the advantages of the distributed approach by showing the almost linear decreasing of average training time according to the number of processors. In a future work, we expect to apply this methodology to mixture-of-experts with embedded variable selection.
AB - Today, organizations are beginning to realize the importance of using as much data as possible for decision-making in their strategy. The finding of relevant patterns in enormous amount of data requires automatic machine learning algorithms, among them, a popular option is the mixture-of-experts that allows to model data using a set of local experts. The problem of using typical learning algorithms over Big Data is the handling of these large datasets in primary memory. In this paper, we propose a methodology to learn a mixture-of-experts in a distributed way using PETUUM platform. Particularly, we propose to learn the parameters of mixture-of-experts by adapting the standard stochastic gradient descent in a distributed way. This methodology is applied to people detection with standard real datasets considering accuracy and precision metrics among other. The results show a consistent performance of mixture-of-experts models where the best number of experts varies according to the particular dataset. We also evidence the advantages of the distributed approach by showing the almost linear decreasing of average training time according to the number of processors. In a future work, we expect to apply this methodology to mixture-of-experts with embedded variable selection.
UR - http://www.scopus.com/inward/record.url?scp=85050966719&partnerID=8YFLogxK
U2 - 10.1109/SCCC.2017.8405113
DO - 10.1109/SCCC.2017.8405113
M3 - Conference contribution
AN - SCOPUS:85050966719
VL - 2017-October
T3 - Proceedings - International Conference of the Chilean Computer Science Society, SCCC
SP - 1
EP - 7
BT - 2017 36th International Conference of the Chilean Computer Science Society, SCCC 2017
PB - IEEE Computer Society
T2 - 36th International Conference of the Chilean Computer Science Society, SCCC 2017
Y2 - 16 October 2017 through 20 October 2017
ER -