Skip to content

Latest commit

 

History

History
executable file
·
80 lines (75 loc) · 3.4 KB

ReadMe.md

File metadata and controls

executable file
·
80 lines (75 loc) · 3.4 KB

Overview

This repository includes implementations of speaker verification systems that input raw waveforms.

Currently, it includes four systems in python. Detailed instructions on each system is described in individual ReadME files.

RawNet3

  • PyTorch implementation
  • Performance
    • supervised learning with AAM-Softmax: EER 0.89%
    • self-supervised learning: EER 5.40%
  • Training recipe
  • Inference
    • Pre-trained weight parameters are stored in HuggingFace and is included as a submodule.
    • Vox1-O benchmark is available in RawNet3.
    • Extracting speaker embedding from any 16k 16bit mono utterance is supported.
  • Published as a conference paper in Interspeech 2022.
@article{jung2022pushing,
  title={Pushing the limits of raw waveform speaker recognition},
  author={Jung, Jee-weon and Kim, You Jin and Heo, Hee-Soo and Lee, Bong-Jin and Kwon, Youngki and Chung, Joon Son},
  journal={Proc. Interspeech},
  year={2022}
}

RawNet2_modified

  • Code refactoring
  • Performance
    • EER 1.91%
      • Trained using VoxCeleb2
      • VoxCeleb1 original trial
    • Will be used as a baseline system for authors' future works

RawNet2

@article{jung2020improved,
  title={Improved RawNet with Feature Map Scaling for Text-independent Speaker Verification using Raw Waveforms},
  author={Jung, Jee-weon and Kim, Seung-bin and Shim, Hye-jin and Kim, Ju-ho and Yu, Ha-Jin},
  journal={Proc. Interspeech},
  pages={3583--3587},
  year={2020}
}

RawNet

@article{jung2019RawNet,
  title={RawNet: Advanced end-to-end deep neural network using raw waveforms for text-independent speaker verification},
  author={Jung, Jee-weon and Heo, Hee-soo and Kim, ju-ho and Shim, Hye-jin and Yu, Ha-jin},
  journal={Proc. Interspeech},
  pages={1268--1272},
  year={2019}
}