Skip to content

Hyperparameters are crucial in machine learning, as they determine if the model will be efficient and accurate. Usually, a tuning process is used to find good combinations of hyperparameters. This tuning process used to be very time-consuming as it historically consisted of training the model with all combinations of hyperparameters among the sp…

Notifications You must be signed in to change notification settings

AJSVB/Novel-Approaches-For-Tuning-Models-HyperParameters

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

31 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Novel Approaches For Tuning Models HyperParameters (Optimization)

Hyperparameters are crucial in machine learning, as they determine if the model will be efficient and accurate. Usually, a tuning process is used to find good combinations of hyperparameters. This tuning process used to be very time-consuming as it historically consisted of training the model with all combinations of hyperparameters among the space of all possibilities for grid-search, and of a random set of combinations for random search. Then results were compared to choose the best combination. Now notice that the space of possibilities grows exponentially with the number of hyperparameters. Furthermore, for complex deep-learning problems a single training can take days. Thus new approaches are needed to find good hyperparameters faster. In this project, we would like to choose a few deep-learning problems as they are usually long to train and compare different tuning approaches using the Ray Tune library. We would like to empirically quantify and qualify the speed for each method as well as the efficiency of the parameters found.

Please look at "SEMESTER_PROJECT(5)" to see the last version of the semster project.

Grade: 6/6 !!!

Version: Working with:

   !pip install ray==1.0.0,
    
   !pip install ax-platform
   
   !pip install sqlalchemy==1.3.19  
    
   !pip install nevergrad==0.4.2
   
   !pip install configspace==0.4.15
   
   !pip install zoopt==0.4.0
   
   !pip install adabelief-pytorch==0.1.0
   
   !pip install tensorboardX==2.1
   
   !pip install HpBandSter==0.7.4

from a google collab environment (see IMDB in first batch)

About

Hyperparameters are crucial in machine learning, as they determine if the model will be efficient and accurate. Usually, a tuning process is used to find good combinations of hyperparameters. This tuning process used to be very time-consuming as it historically consisted of training the model with all combinations of hyperparameters among the sp…

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published