Mostra el registre d'ítem simple

dc.contributor.authorKhan, Umair
dc.contributor.authorHernando Pericás, Francisco Javier
dc.contributor.otherUniversitat Politècnica de Catalunya. Doctorat en Teoria del Senyal i Comunicacions
dc.contributor.otherUniversitat Politècnica de Catalunya. Departament de Teoria del Senyal i Comunicacions
dc.date.accessioned2020-11-12T17:06:10Z
dc.date.available2020-11-12T17:06:10Z
dc.date.issued2020
dc.identifier.citationKhan, U.; Hernando, J. Unsupervised training of siamese networks for speaker verification. A: Annual Conference of the International Speech Communication Association. "Interspeech 2020: the 20th Annual Conference of the International Speech Communication Association: 25-29 October 2020: Shanghai, China". Baixas: International Speech Communication Association (ISCA), 2020, p. 3002-3006. ISBN 1990-9772. DOI 10.21437/Interspeech.2020-1882.
dc.identifier.isbn1990-9772
dc.identifier.urihttp://hdl.handle.net/2117/332092
dc.description.abstractSpeaker labeled background data is an essential requirement for most state-of-the-art approaches in speaker recognition, e.g., xvectors and i-vector/PLDA. However, in reality it is difficult to access large amount of labeled data. In this work, we propose siamese networks for speaker verification without using speaker labels. We propose two different siamese networks having two and three branches, respectively, where each branch is a CNN encoder. Since the goal is to avoid speaker labels, we propose to generate the training pairs in an unsupervised manner. The client samples are selected within one database according to highest cosine scores with the anchor in i-vector space. The impostor samples are selected in the same way but from another database. Our double-branch siamese performs binary classification using cross entropy loss during training. In testing phase, we obtain speaker verification scores directly from its output layer. Whereas, our triple-branch siamese is trained to learn speaker embeddings using triplet loss. During testing, we extract speaker embeddings from its output layer, which are scored in the experiments using cosine scoring. The evaluation is performed on VoxCeleb-1 database, which show that using the proposed unsupervised systems, solely or in fusion, the results get closer to supervised baseline
dc.description.sponsorshipThis work has been developed in the framework of DeepVoice Project (TEC2015-69266-P), funded by Spanish Ministry
dc.format.extent5 p.
dc.language.isoeng
dc.publisherInternational Speech Communication Association (ISCA)
dc.subjectÀrees temàtiques de la UPC::Enginyeria de la telecomunicació::Processament del senyal::Processament de la parla i del senyal acústic
dc.subject.lcshAutomatic speech recognition
dc.subject.otheri-vector
dc.subject.otherImpostor selection
dc.subject.otherCNN
dc.subject.otherTriplet loss
dc.titleUnsupervised training of siamese networks for speaker verification
dc.typeConference report
dc.subject.lemacReconeixement automàtic de la parla
dc.contributor.groupUniversitat Politècnica de Catalunya. VEU - Grup de Tractament de la Parla
dc.identifier.doi10.21437/Interspeech.2020-1882
dc.description.peerreviewedPeer Reviewed
dc.relation.publisherversionhttp://dx.doi.org/10.21437/Interspeech.2020-1882
dc.rights.accessOpen Access
local.identifier.drac29753620
dc.description.versionPostprint (published version)
dc.relation.projectidinfo:eu-repo/grantAgreement/MINECO//TEC2015-69266-P/ES/TECNOLOGIAS DE APRENDIZAJE PROFUNDO APLICADAS AL PROCESADO DE VOZ Y AUDIO/
local.citation.authorKhan, U.; Hernando, J.
local.citation.contributorAnnual Conference of the International Speech Communication Association
local.citation.pubplaceBaixas
local.citation.publicationNameInterspeech 2020: the 20th Annual Conference of the International Speech Communication Association: 25-29 October 2020: Shanghai, China
local.citation.startingPage3002
local.citation.endingPage3006


Fitxers d'aquest items

Thumbnail

Aquest ítem apareix a les col·leccions següents

Mostra el registre d'ítem simple