import sys,os,multiprocessing now_dir=os.getcwd() sys.path.append(now_dir) inp_root = sys.argv[1] sr = int(sys.argv[2]) n_p = int(sys.argv[3]) exp_dir = sys.argv[4] noparallel = sys.argv[5] == "True" import numpy as np,os,traceback from slicer2 import Slicer import librosa,traceback from scipy.io import wavfile import multiprocessing from my_utils import load_audio mutex = multiprocessing.Lock() class PreProcess(): def __init__(self,sr,exp_dir): self.slicer = Slicer( sr=sr, threshold=-32, min_length=800, min_interval=400, hop_size=15, max_sil_kept=150 ) self.sr=sr self.per=3.7 self.overlap=0.3 self.tail=self.per+self.overlap self.max=0.95 self.alpha=0.8 self.exp_dir=exp_dir self.gt_wavs_dir="%s/0_gt_wavs"%exp_dir self.wavs16k_dir="%s/1_16k_wavs"%exp_dir self.f = open("%s/preprocess.log"%exp_dir, "a+") os.makedirs(self.exp_dir,exist_ok=True) os.makedirs(self.gt_wavs_dir,exist_ok=True) os.makedirs(self.wavs16k_dir,exist_ok=True) def print(self, strr): mutex.acquire() print(strr) self.f.write("%s\n" % strr) self.f.flush() mutex.release() def norm_write(self,tmp_audio,idx0,idx1): tmp_audio = (tmp_audio / np.abs(tmp_audio).max() * (self.max * self.alpha)) + (1 - self.alpha) * tmp_audio wavfile.write("%s/%s_%s.wav" % (self.gt_wavs_dir, idx0, idx1), self.sr, (tmp_audio*32768).astype(np.int16)) tmp_audio = librosa.resample(tmp_audio, orig_sr=self.sr, target_sr=16000) wavfile.write("%s/%s_%s.wav" % (self.wavs16k_dir, idx0, idx1), 16000, (tmp_audio*32768).astype(np.int16)) def pipeline(self,path, idx0): try: audio = load_audio(path,self.sr) idx1=0 for audio in self.slicer.slice(audio): i = 0 while (1): start = int(self.sr * (self.per - self.overlap) * i) i += 1 if (len(audio[start:]) > self.tail * self.sr): tmp_audio = audio[start:start + int(self.per * self.sr)] self.norm_write(tmp_audio,idx0,idx1) idx1 += 1 else: tmp_audio = audio[start:] break self.norm_write(tmp_audio, idx0, idx1) self.print("%s->Suc."%path) except: self.print("%s->%s"%(path,traceback.format_exc())) def pipeline_mp(self,infos): for path, idx0 in infos: self.pipeline(path,idx0) def pipeline_mp_inp_dir(self,inp_root,n_p): try: infos = [("%s/%s" % (inp_root, name), idx) for idx, name in enumerate(sorted(list(os.listdir(inp_root))))] if noparallel: for i in range(n_p): self.pipeline_mp(infos[i::n_p]) else: ps=[] for i in range(n_p): p=multiprocessing.Process(target=self.pipeline_mp,args=(infos[i::n_p],)) p.start() ps.append(p) for p in ps:p.join() except: self.print("Fail. %s"%traceback.format_exc()) def preprocess_trainset(inp_root, sr, n_p, exp_dir): pp=PreProcess(sr,exp_dir) pp.print("start preprocess") pp.print(sys.argv) pp.pipeline_mp_inp_dir(inp_root,n_p) pp.print("end preprocess") if __name__=='__main__': preprocess_trainset(inp_root, sr, n_p, exp_dir)