#!/usr/bin/env python3 # Copyright 2023 Xiaomi Corp. (authors: Fangjun Kuang) # Please see ./run.sh for usage from typing import Optional import ffmpeg import numpy as np import soundfile as sf import torch from pydub import AudioSegment from unet import UNet def load_audio(filename, sample_rate: Optional[int] = 44100): probe = ffmpeg.probe(filename) if "streams" not in probe or len(probe["streams"]) == 0: raise ValueError("No stream was found with ffprobe") metadata = next( stream for stream in probe["streams"] if stream["codec_type"] == "audio" ) n_channels = metadata["channels"] if sample_rate is None: sample_rate = metadata["sample_rate"] process = ( ffmpeg.input(filename) .output("pipe:", format="f32le", ar=sample_rate) .run_async(pipe_stdout=True, pipe_stderr=True) ) buffer, _ = process.communicate() waveform = np.frombuffer(buffer, dtype=" 2: waveform = waveform[:, :2] return waveform, sample_rate @torch.no_grad() def main(): vocals = UNet() vocals.eval() state_dict = torch.load("./2stems/vocals.pt", map_location="cpu") vocals.load_state_dict(state_dict) accompaniment = UNet() accompaniment.eval() state_dict = torch.load("./2stems/accompaniment.pt", map_location="cpu") accompaniment.load_state_dict(state_dict) # # waveform, sample_rate = load_audio("./audio_example.mp3") # You can download the following two mp3 from # https://huggingface.co/spaces/csukuangfj/music-source-separation/tree/main/examples waveform, sample_rate = load_audio("./qi-feng-le.mp3") # waveform, sample_rate = load_audio("./Yesterday_Once_More-Carpenters.mp3") assert waveform.shape[1] == 2, waveform.shape waveform = torch.nn.functional.pad(waveform, (0, 0, 0, 4096)) # torch.stft requires a 2-D input of shape (N, T), so we transpose waveform stft = torch.stft( waveform.t(), n_fft=4096, hop_length=1024, window=torch.hann_window(4096, periodic=True), center=False, onesided=True, return_complex=True, ) print("stft", stft.shape) # stft: (2, 2049, 465) # stft is a complex tensor y = stft.permute(2, 1, 0) print("y0", y.shape) # (465, 2049, 2) y = y[:, :1024, :] # (465, 1024, 2) tensor_size = y.shape[0] - int(y.shape[0] / 512) * 512 pad_size = 512 - tensor_size y = torch.nn.functional.pad(y, (0, 0, 0, 0, 0, pad_size)) # (512, 1024, 2) print("y1", y.shape, y.dtype) num_splits = int(y.shape[0] / 512) y = y.reshape([num_splits, 512] + list(y.shape[1:])) # y: (1, 512, 1024, 2) print("y2", y.shape, y.dtype) y = y.abs() y = y.permute(3, 0, 1, 2) # (2, 1, 512, 1024) print("y3", y.shape, y.dtype) vocals_spec = vocals(y) accompaniment_spec = accompaniment(y) vocals_spec = vocals_spec.permute(1, 0, 2, 3) accompaniment_spec = accompaniment_spec.permute(1, 0, 2, 3) sum_spec = (vocals_spec**2 + accompaniment_spec**2) + 1e-10 print( "vocals_spec", vocals_spec.shape, accompaniment_spec.shape, sum_spec.shape, vocals_spec.dtype, ) vocals_spec = (vocals_spec**2 + 1e-10 / 2) / sum_spec # (1, 2, 512, 1024) accompaniment_spec = (accompaniment_spec**2 + 1e-10 / 2) / sum_spec # (1, 2, 512, 1024) for name, spec in zip( ["vocals", "accompaniment"], [vocals_spec, accompaniment_spec] ): spec = torch.nn.functional.pad(spec, (0, 2049 - 1024, 0, 0, 0, 0, 0, 0)) # (1, 2, 512, 2049) spec = spec.permute(0, 2, 3, 1) # (1, 512, 2049, 2) print("here00", spec.shape) spec = spec.reshape(-1, spec.shape[2], spec.shape[3]) # (512, 2049, 2) print("here2", spec.shape) # (512, 2049, 2) spec = spec[: stft.shape[2], :, :] # (465, 2049, 2) print("here 3", spec.shape, stft.shape) spec = spec.permute(2, 1, 0) # (2, 2049, 465) masked_stft = spec * stft wave = torch.istft( masked_stft, 4096, 1024, window=torch.hann_window(4096, periodic=True), onesided=True, ) * (2 / 3) print(wave.shape, wave.dtype) sf.write(f"{name}.wav", wave.t(), 44100) wave = (wave.t() * 32768).to(torch.int16) sound = AudioSegment( data=wave.numpy().tobytes(), sample_width=2, frame_rate=44100, channels=2 ) sound.export(f"{name}.mp3", format="mp3", bitrate="128k") if __name__ == "__main__": main()