From e87e3c5ac01ac3d154dca8496c5e73783a358742 Mon Sep 17 00:00:00 2001 From: Gustaf Rydholm Date: Sat, 27 Nov 2021 12:37:18 +0100 Subject: Update lines config with no axial encoder --- training/conf/experiment/conv_transformer_lines.yaml | 20 ++++++++++---------- 1 file changed, 10 insertions(+), 10 deletions(-) (limited to 'training/conf/experiment') diff --git a/training/conf/experiment/conv_transformer_lines.yaml b/training/conf/experiment/conv_transformer_lines.yaml index 3af2c37..38b72c4 100644 --- a/training/conf/experiment/conv_transformer_lines.yaml +++ b/training/conf/experiment/conv_transformer_lines.yaml @@ -64,7 +64,7 @@ rotary_embedding: &rotary_embedding dim: 64 attn: &attn - dim: &hidden_dim 256 + dim: &hidden_dim 128 num_heads: 4 dim_head: 64 dropout_rate: &dropout_rate 0.2 @@ -76,7 +76,7 @@ network: num_classes: *num_classes pad_index: *ignore_index encoder: - _target_: text_recognizer.networks.efficientnet.EfficientNet + _target_: text_recognizer.networks.efficientnet.efficientnet.EfficientNet arch: b0 depth: 5 stochastic_dropout_rate: 0.2 @@ -109,14 +109,14 @@ network: _target_: text_recognizer.networks.transformer.embeddings.axial.AxialPositionalEmbedding dim: *hidden_dim shape: &shape [3, 64] - axial_encoder: - _target_: text_recognizer.networks.transformer.axial_attention.encoder.AxialEncoder - dim: *hidden_dim - heads: 4 - shape: *shape - depth: 2 - dim_head: 64 - dim_index: 1 + axial_encoder: null + # _target_: text_recognizer.networks.transformer.axial_attention.encoder.AxialEncoder + # dim: *hidden_dim + # heads: 4 + # shape: *shape + # depth: 2 + # dim_head: 64 + # dim_index: 1 model: _target_: text_recognizer.models.transformer.TransformerLitModel -- cgit v1.2.3-70-g09d2