summaryrefslogtreecommitdiff
path: root/text_recognizer/networks
AgeCommit message (Expand)Author
2021-11-27Revert "Remove default transformer"Gustaf Rydholm
2021-11-22Fix import efficientnetGustaf Rydholm
2021-11-22Remove unused utilsGustaf Rydholm
2021-11-22Format conv transformerGustaf Rydholm
2021-11-22Move efficientnet from encoder dirGustaf Rydholm
2021-11-21Format filesGustaf Rydholm
2021-11-21Add axial encoder to conv transformerGustaf Rydholm
2021-11-21Remove VQVAE stuff, did not work...Gustaf Rydholm
2021-11-21Add axial transformerGustaf Rydholm
2021-11-17Update VqTransformerGustaf Rydholm
2021-11-17Update encoder fun in conv_transformerGustaf Rydholm
2021-11-17Update vqvae with new quantizerGustaf Rydholm
2021-11-17Add new quantizerGustaf Rydholm
2021-11-17Remove local attentionGustaf Rydholm
2021-11-17Replace last two layers in efficientnet with oneGustaf Rydholm
2021-11-17Format efficientnetGustaf Rydholm
2021-11-07Fix dropout in efficientnetGustaf Rydholm
2021-11-07Add dropout to final layer in efficientnetGustaf Rydholm
2021-11-07Fix mbconv sub modulesGustaf Rydholm
2021-11-06Fix efficientnet incorrect channel calculationGustaf Rydholm
2021-11-06Format efficientnetGustaf Rydholm
2021-11-05Remove conv attentionGustaf Rydholm
2021-11-05Rename mask to input_maskGustaf Rydholm
2021-11-05Remove out_channels as a settable parameter in effnetGustaf Rydholm
2021-11-03Remove unused args from conv transformerGustaf Rydholm
2021-11-03Remove pixelcnnGustaf Rydholm
2021-11-03Fix output from attn modulesGustaf Rydholm
2021-11-03Fix local attn to work with any input lengthGustaf Rydholm
2021-11-03Update output shape from attn moduleGustaf Rydholm
2021-11-01Add check for positional encoding in attn layerGustaf Rydholm
2021-11-01Fix bugs in local attentionGustaf Rydholm
2021-11-01Refactor transformer layerGustaf Rydholm
2021-11-01Fix self attention moduleGustaf Rydholm
2021-11-01Update rotary embeddingGustaf Rydholm
2021-10-28Update rotary embeddingGustaf Rydholm
2021-10-28Remove 2D positional embeddingGustaf Rydholm
2021-10-28Remove absolute positional embeddingGustaf Rydholm
2021-10-28Add check for position embeddingGustaf Rydholm
2021-10-28Fix multihead local attentionGustaf Rydholm
2021-10-28Fix bug with local attentionGustaf Rydholm
2021-10-28Refactor attention moduleGustaf Rydholm
2021-10-27Remove Barlow TwinsGustaf Rydholm
2021-10-27Add axial embeddingGustaf Rydholm
2021-10-27Remove ViTGustaf Rydholm
2021-10-27Remove default transformerGustaf Rydholm
2021-10-27Add comments to transformer modulesGustaf Rydholm
2021-10-27Clean up local attention, add comments and typesGustaf Rydholm
2021-10-27Add local attn in transformer layerGustaf Rydholm
2021-10-27Remove unused import and add comments in attn moduleGustaf Rydholm
2021-10-27Remove imports from __init__ in transformer networkGustaf Rydholm