Skip to main content
SHARE
Publication

Optimizing deep learning hyper-parameters through an evolutionary algorithm

Publication Type
Journal
Journal Name
MLHPC '15 Proceedings of the Workshop on Machine Learning in High-Performance Computing Environments
Publication Date

There has been a recent surge of success in utilizing Deep Learning (DL) in imaging and speech applications for its relatively automatic feature generation and, in particular for convolutional neural networks (CNNs), high accuracy classification abilities. While these models learn their parameters through data-driven methods, model selection (as architecture construction) through hyper-parameter choices remains a tedious and highly intuition driven task. To address this, Multi-node Evolutionary Neural Networks for Deep Learning (MENNDL) is proposed as a method for automating network selection on computational clusters through hyper-parameter optimization performed via genetic algorithms.