diff options
Diffstat (limited to 'src/training/experiments')
-rw-r--r-- | src/training/experiments/sample_experiment.yml | 96 |
1 files changed, 84 insertions, 12 deletions
diff --git a/src/training/experiments/sample_experiment.yml b/src/training/experiments/sample_experiment.yml index 70edb63..57198f1 100644 --- a/src/training/experiments/sample_experiment.yml +++ b/src/training/experiments/sample_experiment.yml @@ -1,28 +1,30 @@ experiment_group: Sample Experiments experiments: - - dataloader: EmnistDataLoaders - data_loader_args: - splits: [train, val] + - dataset: EmnistDataset + dataset_args: sample_to_balance: true subsample_fraction: null transform: null target_transform: null + seed: 4711 + data_loader_args: + splits: [train, val] batch_size: 256 shuffle: true num_workers: 8 cuda: true - seed: 4711 model: CharacterModel metrics: [accuracy] - # network: MLP - # network_args: - # input_size: 784 - # output_size: 62 - # num_layers: 3 - network: LeNet + network: MLP network_args: - input_size: [28, 28] + input_size: 784 output_size: 62 + num_layers: 3 + activation_fn: GELU + # network: LeNet + # network_args: + # output_size: 62 + # activation_fn: GELU train_args: batch_size: 256 epochs: 16 @@ -66,5 +68,75 @@ experiments: num_examples: 4 OneCycleLR: null - verbosity: 1 # 0, 1, 2 + verbosity: 2 # 0, 1, 2 resume_experiment: null + # - dataset: EmnistDataset + # dataset_args: + # sample_to_balance: true + # subsample_fraction: null + # transform: null + # target_transform: null + # seed: 4711 + # data_loader_args: + # splits: [train, val] + # batch_size: 256 + # shuffle: true + # num_workers: 8 + # cuda: true + # model: CharacterModel + # metrics: [accuracy] + # # network: MLP + # # network_args: + # # input_size: 784 + # # output_size: 62 + # # num_layers: 3 + # # activation_fn: GELU + # network: LeNet + # network_args: + # output_size: 62 + # activation_fn: GELU + # train_args: + # batch_size: 256 + # epochs: 16 + # criterion: CrossEntropyLoss + # criterion_args: + # weight: null + # ignore_index: -100 + # reduction: mean + # # optimizer: RMSprop + # # optimizer_args: + # # lr: 1.e-3 + # # alpha: 0.9 + # # eps: 1.e-7 + # # momentum: 0 + # # weight_decay: 0 + # # centered: false + # optimizer: AdamW + # optimizer_args: + # lr: 1.e-2 + # betas: [0.9, 0.999] + # eps: 1.e-08 + # weight_decay: 0 + # amsgrad: false + # # lr_scheduler: null + # lr_scheduler: OneCycleLR + # lr_scheduler_args: + # max_lr: 1.e-3 + # epochs: 16 + # callbacks: [Checkpoint, EarlyStopping, WandbCallback, WandbImageLogger, OneCycleLR] + # callback_args: + # Checkpoint: + # monitor: val_accuracy + # EarlyStopping: + # monitor: val_loss + # min_delta: 0.0 + # patience: 3 + # mode: min + # WandbCallback: + # log_batch_frequency: 10 + # WandbImageLogger: + # num_examples: 4 + # OneCycleLR: + # null + # verbosity: 2 # 0, 1, 2 + # resume_experiment: null |