From fb90a53b1235fd836dee74452f3f2a621e0f363a Mon Sep 17 00:00:00 2001 From: Gustaf Rydholm Date: Wed, 27 Oct 2021 22:13:54 +0200 Subject: Rename transformer embeddings --- .../positional_encodings/absolute_embedding.py | 17 ----------------- 1 file changed, 17 deletions(-) delete mode 100644 text_recognizer/networks/transformer/positional_encodings/absolute_embedding.py (limited to 'text_recognizer/networks/transformer/positional_encodings/absolute_embedding.py') diff --git a/text_recognizer/networks/transformer/positional_encodings/absolute_embedding.py b/text_recognizer/networks/transformer/positional_encodings/absolute_embedding.py deleted file mode 100644 index 7140537..0000000 --- a/text_recognizer/networks/transformer/positional_encodings/absolute_embedding.py +++ /dev/null @@ -1,17 +0,0 @@ -"""Absolute positional embedding.""" -import torch -from torch import nn, Tensor - - -class AbsolutePositionalEmbedding(nn.Module): - def __init__(self, dim: int, max_seq_len: int) -> None: - super().__init__() - self.emb = nn.Embedding(max_seq_len, dim) - self._weight_init() - - def _weight_init(self) -> None: - nn.init.normal_(self.emb.weight, std=0.02) - - def forward(self, x: Tensor) -> Tensor: - n = torch.arange(x.shape[1], device=x.device) - return self.emb(n)[None, :, :] -- cgit v1.2.3-70-g09d2