From 732e4a152d29842d5021d810a427d2ad13231c75 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Thu, 11 Apr 2024 16:58:06 +0900 Subject: [PATCH] chore(format): run black on dev (#1870) Co-authored-by: github-actions[bot] --- infer/lib/infer_pack/models.py | 16 +++++++++++----- 1 file changed, 11 insertions(+), 5 deletions(-) diff --git a/infer/lib/infer_pack/models.py b/infer/lib/infer_pack/models.py index 7908c23..262814d 100644 --- a/infer/lib/infer_pack/models.py +++ b/infer/lib/infer_pack/models.py @@ -52,7 +52,13 @@ class TextEncoder(nn.Module): ) self.proj = nn.Conv1d(hidden_channels, out_channels * 2, 1) - def forward(self, phone: torch.Tensor, pitch: torch.Tensor, lengths: torch.Tensor, skip_head: Optional[torch.Tensor] = None): + def forward( + self, + phone: torch.Tensor, + pitch: torch.Tensor, + lengths: torch.Tensor, + skip_head: Optional[torch.Tensor] = None, + ): if pitch is None: x = self.emb_phone(phone) else: @@ -67,8 +73,8 @@ class TextEncoder(nn.Module): if skip_head is not None: assert isinstance(skip_head, torch.Tensor) head = int(skip_head.item()) - x = x[:, :, head : ] - x_mask = x_mask[:, :, head : ] + x = x[:, :, head:] + x_mask = x_mask[:, :, head:] stats = self.proj(x) * x_mask m, logs = torch.split(stats, self.out_channels, dim=1) return m, logs, x_mask @@ -807,7 +813,7 @@ class SynthesizerTrnMs768NSFsid(SynthesizerTrnMs256NSFsid): spk_embed_dim, gin_channels, sr, - **kwargs + **kwargs ) del self.enc_p self.enc_p = TextEncoder( @@ -1021,7 +1027,7 @@ class SynthesizerTrnMs768NSFsid_nono(SynthesizerTrnMs256NSFsid_nono): spk_embed_dim, gin_channels, sr, - **kwargs + **kwargs ) del self.enc_p self.enc_p = TextEncoder(