Need help with VGG-Speaker-Recognition?
Click the “chat” button below for chat support from the developer who created it, or find similar developers for support.

About the developer

WeidiXie
239 Stars 80 Forks 41 Commits 2 Opened issues

Description

Utterance-level Aggregation For Speaker Recognition In The Wild

Services available

!
?

Need anything else?

Contributors list

No Data

README

This repo will contain the code for ICASSP 2019, speaker identifcation (http://www.robots.ox.ac.uk/~vgg/research/speakerID/).

This repo contains a Keras implementation of the paper,
Utterance-level Aggregation For Speaker Recognition In The Wild (Xie et al., ICASSP 2019) (Oral).

**New challenge on speaker recognition: The VoxCeleb Speaker Recognition Challenge (VoxSRC).

Dependencies

Data

The dataset used for the experiments are

Training the model

To train the model on the Voxceleb2 dataset, you can run

  • python main.py --net resnet34s --batchsize 160 --gpu 2,3 --lr 0.001 --warmupratio 0.1 --optimizer adam --epochs 128 --multiprocess 8 --loss softmax --datapath ../pathto_voxceleb2

Model

  • Models:

    google drive: https://drive.google.com/open?id=1M_SXoW1ceKm3LghItY2ENKKUn3cWYfZm

    dropbox: https://www.dropbox.com/sh/n96ekf7ilsvkjdp/AACXKDesS2ju5rp6Cyxh2PCva?dl=0

  • Download the models and put it in the folder, model/

Testing the model

To test a specific model on the voxceleb1 dataset, for example, the model trained with ResNet34s trained by adam with softmax, and feature dimension 512

  • python predict.py --gpu 1 --net resnet34s --ghostcluster 2 --vladcluster 8 --loss softmax --resume ../model/gvladsoftmax/resnet34vlad8ghost2bdim512_deploy/weights.h5

  • Results:

    VoxCeleb1-Test: 3.22        VoxCeleb1-Test-Cleaned: 3.24
    VoxCeleb1-Test-E: 3.24      VoxCeleb1-Test-E-Cleaned: 3.13
    VoxCeleb1-Test-H: 5.17      VoxCeleb1-Test-H-Cleaned: 5.06
    

Licence

The code and mode are available to download for commercial/research purposes under a Creative Commons Attribution 4.0 International License(https://creativecommons.org/licenses/by/4.0/).

  Downloading this code implies agreement to follow the same conditions for any modification 
  and/or re-distribution of the dataset in any form.

Additionally any entity using this code agrees to the following conditions:

THIS CODE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.

Please cite the papers below if you make use of the dataset and code.

Citation

@InProceedings{Xie19,
  author       = "W. Xie, A. Nagrani, J. S. Chung, A. Zisserman",
  title        = "Utterance-level Aggregation For Speaker Recognition In The Wild.",
  booktitle    = "ICASSP, 2019",
  year         = "2019",
}

@Article{Nagrani19, author = "A. Nagrani, J. S. Chung, W. Xie, A. Zisserman", title = "VoxCeleb: Large-scale Speaker Verification in the Wild.", journal = "Computer Speech & Language", year = "2019", }

We use cookies. If you continue to browse the site, you agree to the use of cookies. For more information on our use of cookies please see our Privacy Policy.