Switch to side-by-side view

--- a
+++ b/5-Training with Ignite and Optuna/constants.py
@@ -0,0 +1,25 @@
+import optuna
+
+MODEL_NAME='MyModel' #the name of the class in models.py you want to tune
+OPTIMIZERS=['AdamW', 'RMSprop'] #the list of optimizers to sample from
+MAX_EPOCHS=100 #the max number of epochs to train for each hyperparameter combination
+METRIC='roc_auc' #the metric to optimize for across trials, also used for early stopping within a trial and pruning across trials, full list of possible values in tuningfunctions.Objective
+MIN_LR=3e-5 #the mininum learning rate to sample from on a log scale
+MAX_LR=3e-3 #the maximum learning rate to sample from on a log scale
+PATIENCE=10 #the early stopping patience, can be set to None
+SCHEDULER=False #bool for whether to use the learning rate scheduler
+STEP=1 #the learning rate scheduler step size, if SCHEDULER=True, cannot be None
+GAMMA=0.975 #the learning rate gamma, if SHCEDULER=True, cannot be None
+
+PRUNER=optuna.pruners.NopPruner() #the optuna pruner to use across trials
+NUM_TRIALS=5 #the number of different hyperparameter combination to try
+DIRECTION='maximize' #the direction of the metric to optimize towards ex: 'loss' = 'minimize', 'roc_auc' = 'maximize'
+
+
+
+## MyModel specific parameters
+INITIAL_KERNEL_NUM = [4,8,16,32,64]
+MIN_DROPOUT = 0
+MAX_DROPOUT = 1
+CONV1_KERNEL1 = [7,21]
+CONV1_KERNEL2 = [1,3]
\ No newline at end of file