Master of Computing and Information Systems, Youngstown State University, 2023, Department of Computer Science and Information Systems
The performance of machine learning (ML) methods, including deep learning, for classification and regression tasks applied to tabular datasets is sensitive to hyperparameters values. Therefore, finding the optimal values of these hyperparameters is integral to improving the prediction accuracy of a machine learning algorithm and the model selection. However, manually searching for the best configuration is a tedious task, and many AutoML (automated machine learning) frameworks have been proposed recently to help practitioners solve this problem. Hyperparameters are the values or configurations used to control the algorithm's behavior while building the model. Hyperparameter optimization is the guided process of finding the best combination of hyperparameters that delivers the best performance on the data and task at hand in a reasonable amount of time. In this work, the performance of two frequently used AutoML hyperparameter optimization frameworks, Optuna and HyperOpt, are compared on popular OpenML tabular datasets to identify the best framework for tabular data. The results of the experiments show that the performance score of Optuna is better than that of HyperOpt, while HyperOpt is the fastest for hyperparameter optimization.
Committee: Alina Lazar PhD (Advisor); Feng Yu PhD (Committee Member); John R. Sullins PhD (Committee Member)
Subjects: Artificial Intelligence; Comparative; Computer Science; Information Systems