summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorGustaf Rydholm <gustaf.rydholm@gmail.com>2021-12-05 20:27:21 +0100
committerGustaf Rydholm <gustaf.rydholm@gmail.com>2021-12-05 20:27:21 +0100
commit3e516481875c5884cb85ab1baf514567edf483ff (patch)
treef2541319f73c1ceddf1e7a789d07c5c958a69d7c
parent5db1486dd39d123352e529354f47e80c69a89f48 (diff)
Update config lines
-rw-r--r--training/conf/experiment/conv_transformer_lines.yaml26
1 files changed, 13 insertions, 13 deletions
diff --git a/training/conf/experiment/conv_transformer_lines.yaml b/training/conf/experiment/conv_transformer_lines.yaml
index 1905fcf..cca452d 100644
--- a/training/conf/experiment/conv_transformer_lines.yaml
+++ b/training/conf/experiment/conv_transformer_lines.yaml
@@ -10,7 +10,7 @@ defaults:
- override /lr_schedulers: null
- override /optimizers: null
-epochs: &epochs 620
+epochs: &epochs 200
ignore_index: &ignore_index 3
num_classes: &num_classes 57
max_output_len: &max_output_len 89
@@ -28,7 +28,7 @@ callbacks:
stochastic_weight_averaging:
_target_: pytorch_lightning.callbacks.StochasticWeightAveraging
swa_epoch_start: 0.75
- swa_lrs: 1.0e-4
+ swa_lrs: 1.0e-5
annealing_epochs: 10
annealing_strategy: cos
device: null
@@ -36,7 +36,7 @@ callbacks:
optimizers:
madgrad:
_target_: madgrad.MADGRAD
- lr: 3.0e-4
+ lr: 1.0e-4
momentum: 0.9
weight_decay: 0
eps: 1.0e-6
@@ -46,7 +46,7 @@ lr_schedulers:
network:
_target_: torch.optim.lr_scheduler.CosineAnnealingLR
T_max: *epochs
- eta_min: 1.0e-4
+ eta_min: 1.0e-6
last_epoch: -1
interval: epoch
monitor: val/loss
@@ -83,7 +83,7 @@ network:
bn_momentum: 0.99
bn_eps: 1.0e-3
decoder:
- depth: 6
+ depth: 3
_target_: text_recognizer.networks.transformer.layers.Decoder
self_attn:
_target_: text_recognizer.networks.transformer.attention.Attention
@@ -109,14 +109,14 @@ network:
_target_: text_recognizer.networks.transformer.embeddings.axial.AxialPositionalEmbedding
dim: *hidden_dim
shape: &shape [3, 64]
- axial_encoder:
- _target_: text_recognizer.networks.transformer.axial_attention.encoder.AxialEncoder
- dim: *hidden_dim
- heads: 4
- shape: *shape
- depth: 2
- dim_head: 64
- dim_index: 1
+ axial_encoder: null
+ # _target_: text_recognizer.networks.transformer.axial_attention.encoder.AxialEncoder
+ # dim: *hidden_dim
+ # heads: 4
+ # shape: *shape
+ # depth: 2
+ # dim_head: 64
+ # dim_index: 1
model:
_target_: text_recognizer.models.transformer.TransformerLitModel