summaryrefslogtreecommitdiff
path: root/training/experiments/sample_experiment.yml
diff options
context:
space:
mode:
authorGustaf Rydholm <gustaf.rydholm@gmail.com>2021-03-20 18:09:06 +0100
committerGustaf Rydholm <gustaf.rydholm@gmail.com>2021-03-20 18:09:06 +0100
commit7e8e54e84c63171e748bbf09516fd517e6821ace (patch)
tree996093f75a5d488dddf7ea1f159ed343a561ef89 /training/experiments/sample_experiment.yml
parentb0719d84138b6bbe5f04a4982dfca673aea1a368 (diff)
Inital commit for refactoring to lightning
Diffstat (limited to 'training/experiments/sample_experiment.yml')
-rw-r--r--training/experiments/sample_experiment.yml99
1 files changed, 99 insertions, 0 deletions
diff --git a/training/experiments/sample_experiment.yml b/training/experiments/sample_experiment.yml
new file mode 100644
index 0000000..8f94475
--- /dev/null
+++ b/training/experiments/sample_experiment.yml
@@ -0,0 +1,99 @@
+experiment_group: Sample Experiments
+experiments:
+ - train_args:
+ batch_size: 256
+ max_epochs: &max_epochs 32
+ dataset:
+ type: EmnistDataset
+ args:
+ sample_to_balance: true
+ subsample_fraction: null
+ transform: null
+ target_transform: null
+ seed: 4711
+ train_args:
+ num_workers: 6
+ train_fraction: 0.8
+
+ model: CharacterModel
+ metrics: [accuracy]
+ # network: MLP
+ # network_args:
+ # input_size: 784
+ # hidden_size: 512
+ # output_size: 80
+ # num_layers: 5
+ # dropout_rate: 0.2
+ # activation_fn: SELU
+ network:
+ type: ResidualNetwork
+ args:
+ in_channels: 1
+ num_classes: 80
+ depths: [2, 2]
+ block_sizes: [64, 64]
+ activation: leaky_relu
+ # network:
+ # type: WideResidualNetwork
+ # args:
+ # in_channels: 1
+ # num_classes: 80
+ # depth: 10
+ # num_layers: 3
+ # width_factor: 4
+ # dropout_rate: 0.2
+ # activation: SELU
+ # network: LeNet
+ # network_args:
+ # output_size: 62
+ # activation_fn: GELU
+ criterion:
+ type: CrossEntropyLoss
+ args:
+ weight: null
+ ignore_index: -100
+ reduction: mean
+ optimizer:
+ type: AdamW
+ args:
+ lr: 1.e-02
+ betas: [0.9, 0.999]
+ eps: 1.e-08
+ # weight_decay: 5.e-4
+ amsgrad: false
+ # lr_scheduler:
+ # type: OneCycleLR
+ # args:
+ # max_lr: 1.e-03
+ # epochs: *max_epochs
+ # anneal_strategy: linear
+ lr_scheduler:
+ type: CosineAnnealingLR
+ args:
+ T_max: *max_epochs
+ interval: epoch
+ swa_args:
+ start: 2
+ lr: 5.e-2
+ callbacks: [Checkpoint, ProgressBar, WandbCallback, WandbImageLogger, EarlyStopping]
+ callback_args:
+ Checkpoint:
+ monitor: val_accuracy
+ ProgressBar:
+ epochs: null
+ log_batch_frequency: 100
+ EarlyStopping:
+ monitor: val_loss
+ min_delta: 0.0
+ patience: 5
+ mode: min
+ WandbCallback:
+ log_batch_frequency: 10
+ WandbImageLogger:
+ num_examples: 4
+ use_transpose: true
+ verbosity: 0 # 0, 1, 2
+ resume_experiment: null
+ train: true
+ test: true
+ test_metric: test_accuracy