From 9ae5fa1a88899180f88ddb14d4cef457ceb847e5 Mon Sep 17 00:00:00 2001 From: Gustaf Rydholm Date: Mon, 5 Apr 2021 20:47:55 +0200 Subject: Add new training loop with PyTorch Lightning, remove stale files --- training/experiments/embedding_experiment.yml | 64 --------------------------- 1 file changed, 64 deletions(-) delete mode 100644 training/experiments/embedding_experiment.yml (limited to 'training/experiments/embedding_experiment.yml') diff --git a/training/experiments/embedding_experiment.yml b/training/experiments/embedding_experiment.yml deleted file mode 100644 index 1e5f941..0000000 --- a/training/experiments/embedding_experiment.yml +++ /dev/null @@ -1,64 +0,0 @@ -experiment_group: Embedding Experiments -experiments: - - train_args: - transformer_model: false - batch_size: &batch_size 256 - max_epochs: &max_epochs 32 - input_shape: [[1, 28, 28]] - dataset: - type: EmnistDataset - args: - sample_to_balance: true - subsample_fraction: null - transform: null - target_transform: null - seed: 4711 - train_args: - num_workers: 8 - train_fraction: 0.85 - batch_size: *batch_size - model: CharacterModel - metrics: [] - network: - type: DenseNet - args: - growth_rate: 4 - block_config: [4, 4] - in_channels: 1 - base_channels: 24 - num_classes: 128 - bn_size: 4 - dropout_rate: 0.1 - classifier: true - activation: elu - criterion: - type: EmbeddingLoss - args: - margin: 0.2 - type_of_triplets: semihard - optimizer: - type: AdamW - args: - lr: 1.e-02 - betas: [0.9, 0.999] - eps: 1.e-08 - weight_decay: 5.e-4 - amsgrad: false - lr_scheduler: - type: CosineAnnealingLR - args: - T_max: *max_epochs - callbacks: [Checkpoint, ProgressBar, WandbCallback] - callback_args: - Checkpoint: - monitor: val_loss - mode: min - ProgressBar: - epochs: *max_epochs - WandbCallback: - log_batch_frequency: 10 - verbosity: 1 # 0, 1, 2 - resume_experiment: null - train: true - test: true - test_metric: mean_average_precision_at_r -- cgit v1.2.3-70-g09d2