summaryrefslogtreecommitdiff
path: root/text_recognizer/networks/transformer/norm.py
diff options
context:
space:
mode:
authorGustaf Rydholm <gustaf.rydholm@gmail.com>2021-05-01 23:53:50 +0200
committerGustaf Rydholm <gustaf.rydholm@gmail.com>2021-05-01 23:53:50 +0200
commit58ae7154aa945cfe5a46592cc1dfb28f0a4e51b3 (patch)
treec89c1b1a4cc1a499900f2700ab09e8535e2cfe99 /text_recognizer/networks/transformer/norm.py
parent7ae1f8f9654dcea0a9a22310ac0665a5d3202f0f (diff)
Working on new attention module
Diffstat (limited to 'text_recognizer/networks/transformer/norm.py')
-rw-r--r--text_recognizer/networks/transformer/norm.py13
1 files changed, 13 insertions, 0 deletions
diff --git a/text_recognizer/networks/transformer/norm.py b/text_recognizer/networks/transformer/norm.py
index 99a5291..9160876 100644
--- a/text_recognizer/networks/transformer/norm.py
+++ b/text_recognizer/networks/transformer/norm.py
@@ -20,3 +20,16 @@ class Rezero(nn.Module):
def forward(self, x: Tensor, **kwargs: Dict) -> Tensor:
x, *rest = self.fn(x, **kwargs)
return (x * self.g, *rest)
+
+
+class ScaleNorm(nn.Module):
+ def __init__(self, dim: int, eps: float = 1.0e-5) -> None:
+ super().__init__()
+ self.scale = dim ** -0.5
+ self.eps = eps
+ self.g = nn.Parameter(torch.ones(1))
+
+ def forward(self, x: Tensor) -> Tensor:
+ norm = torch.norm(x, dim=-1, keepdim=True) * self.scale
+ return x / norm.clamp(min=self.eps) self.g
+