From 78529eca967207b4d43bf6707fe874cd12d2d405 Mon Sep 17 00:00:00 2001
From: Whi Kwon <30768596+whikwon@users.noreply.github.com>
Date: Wed, 22 Jan 2020 12:18:23 +0900
Subject: [PATCH] Fix typo activatation -> activation (#2007)

---
 mmdet/models/utils/conv_module.py | 6 +++---
 1 file changed, 3 insertions(+), 3 deletions(-)

diff --git a/mmdet/models/utils/conv_module.py b/mmdet/models/utils/conv_module.py
index a1771f8..3be32c3 100644
--- a/mmdet/models/utils/conv_module.py
+++ b/mmdet/models/utils/conv_module.py
@@ -93,7 +93,7 @@ class ConvModule(nn.Module):
         assert set(order) == set(['conv', 'norm', 'act'])
 
         self.with_norm = norm_cfg is not None
-        self.with_activatation = activation is not None
+        self.with_activation = activation is not None
         # if the conv layer is before a norm layer, bias is unnecessary.
         if bias == 'auto':
             bias = False if self.with_norm else True
@@ -135,7 +135,7 @@ class ConvModule(nn.Module):
             self.add_module(self.norm_name, norm)
 
         # build activation layer
-        if self.with_activatation:
+        if self.with_activation:
             # TODO: introduce `act_cfg` and supports more activation layers
             if self.activation not in ['relu']:
                 raise ValueError('{} is currently not supported.'.format(
@@ -162,6 +162,6 @@ class ConvModule(nn.Module):
                 x = self.conv(x)
             elif layer == 'norm' and norm and self.with_norm:
                 x = self.norm(x)
-            elif layer == 'act' and activate and self.with_activatation:
+            elif layer == 'act' and activate and self.with_activation:
                 x = self.activate(x)
         return x
-- 
GitLab