|
import os |
|
|
|
import soundfile as sf |
|
import torch |
|
from librosa.core import load |
|
|
|
from tests import get_tests_input_path, get_tests_output_path, get_tests_path |
|
from TTS.vocoder.layers.pqmf import PQMF |
|
|
|
TESTS_PATH = get_tests_path() |
|
WAV_FILE = os.path.join(get_tests_input_path(), "example_1.wav") |
|
|
|
|
|
def test_pqmf(): |
|
w, sr = load(WAV_FILE) |
|
|
|
layer = PQMF(N=4, taps=62, cutoff=0.15, beta=9.0) |
|
w, sr = load(WAV_FILE) |
|
w2 = torch.from_numpy(w[None, None, :]) |
|
b2 = layer.analysis(w2) |
|
w2_ = layer.synthesis(b2) |
|
|
|
print(w2_.max()) |
|
print(w2_.min()) |
|
print(w2_.mean()) |
|
sf.write(os.path.join(get_tests_output_path(), "pqmf_output.wav"), w2_.flatten().detach(), sr) |
|
|