@@ -51,7 +51,9 @@ def __init__(
5151 cond_layer = torch .nn .Conv1d (
5252 gin_channels , 2 * hidden_channels * n_layers , 1
5353 )
54- self .cond_layer = torch .nn .utils .parametrizations .weight_norm (cond_layer , name = "weight" )
54+ self .cond_layer = torch .nn .utils .parametrizations .weight_norm (
55+ cond_layer , name = "weight"
56+ )
5557
5658 for i in range (n_layers ):
5759 dilation = dilation_rate ** i
@@ -63,7 +65,9 @@ def __init__(
6365 dilation = dilation ,
6466 padding = padding ,
6567 )
66- in_layer = torch .nn .utils .parametrizations .weight_norm (in_layer , name = "weight" )
68+ in_layer = torch .nn .utils .parametrizations .weight_norm (
69+ in_layer , name = "weight"
70+ )
6771 self .in_layers .append (in_layer )
6872
6973 # last one is not necessary
@@ -73,7 +77,9 @@ def __init__(
7377 res_skip_channels = hidden_channels
7478
7579 res_skip_layer = torch .nn .Conv1d (hidden_channels , res_skip_channels , 1 )
76- res_skip_layer = torch .nn .utils .parametrizations .weight_norm (res_skip_layer , name = "weight" )
80+ res_skip_layer = torch .nn .utils .parametrizations .weight_norm (
81+ res_skip_layer , name = "weight"
82+ )
7783 self .res_skip_layers .append (res_skip_layer )
7884
7985 def __call__ (
0 commit comments