From 36875ceca1e00f5bb39a151c50ecf5e333b4cf79 Mon Sep 17 00:00:00 2001 From: Gustaf Rydholm Date: Sun, 21 Nov 2021 23:20:06 +0100 Subject: Update to config --- training/conf/experiment/conv_transformer_lines.yaml | 2 +- training/conf/experiment/conv_transformer_paragraphs.yaml | 11 ++++++++++- 2 files changed, 11 insertions(+), 2 deletions(-) (limited to 'training') diff --git a/training/conf/experiment/conv_transformer_lines.yaml b/training/conf/experiment/conv_transformer_lines.yaml index fe9ef6e..6c266b8 100644 --- a/training/conf/experiment/conv_transformer_lines.yaml +++ b/training/conf/experiment/conv_transformer_lines.yaml @@ -113,7 +113,7 @@ network: dim: *hidden_dim heads: 4 shape: *shape - depth: 2 + depth: 1 dim_head: 64 dim_index: 1 diff --git a/training/conf/experiment/conv_transformer_paragraphs.yaml b/training/conf/experiment/conv_transformer_paragraphs.yaml index d2916e1..4f15ef2 100644 --- a/training/conf/experiment/conv_transformer_paragraphs.yaml +++ b/training/conf/experiment/conv_transformer_paragraphs.yaml @@ -18,6 +18,7 @@ summary: [[1, 1, 576, 640], [1, 682]] criterion: ignore_index: *ignore_index + label_smoothing: 0.1 mapping: &mapping mapping: @@ -108,7 +109,15 @@ network: pixel_pos_embedding: _target_: text_recognizer.networks.transformer.embeddings.axial.AxialPositionalEmbedding dim: *hidden_dim - shape: [18, 20] + shape: &shape [36, 40] + axial_encoder: + _target_: text_recognizer.networks.transformer.axial_attention.encoder.AxialEncoder + dim: *hidden_dim + heads: 4 + shape: *shape + depth: 1 + dim_head: 64 + dim_index: 1 model: _target_: text_recognizer.models.transformer.TransformerLitModel -- cgit v1.2.3-70-g09d2