Experimenting with additive margins for contrastive self-supervised speaker verification
In Proceedings of the 24rd annual conference of the international speech communication association (interspeech 2023)
In Proceedings of the 24rd annual conference of the international speech communication association (interspeech 2023)
In Extraction et gestion des connaissances, EGC 2023, lyon, france, 16 au 20 janvier 2023
In Proceedings of the 23rd annual conference of the international speech communication association (interspeech 2022)
In Workshop EGC 2022 DL for NLP
Hate speech and toxic comment detection on social media has proven to be an essential issue for content moderation. This paper displays a comparison between different Transformer models for Hate Speech detection such as Hate BERT, a BERT-based model, RoBERTa and BERTweet which is a RoBERTa based model. These Transformer models are tested on Jibes&Delight 2021 reddit dataset using the same training and testing conditions. Multiple approaches are detailed in this paper considering feature extraction and data augmentation. The paper concludes that our RoBERTa st4-aug model trained with data augmentation outperforms simple RoBERTa and HateBERT models.
In Computer Speech & Language
In Nature Scientific Reports
In Proc. Interspeech 2019
In NIST speaker recognition evaluation 2016
This document presents the system submission for the group composed of MIT Lincoln Laboratory, Johns Hopkins University (JHU), Laboratoire de Recherche et de Développement de l’EPITA (LRDE) and Universidad Autónoma de Madrid (ATVS). The primary submission is a combination of four systems focused on i-vector systems. Two secondary submissions are also included
In Odyssey 2014, the speaker and language recognition workshop
In this paper, we explored the use of Gaussian Mixture Model (GMM) weights adaptation for speaker verifica- tion. We compared two different subspace weight adap- tation approaches: Subspace Multinomial Model (SMM) and Non-Negative factor Analysis (NFA). Both techniques achieved similar results and seemed to outperform the retraining maximum likelihood (ML) weight adaptation. However, the training process for the NFA approach is substantially faster than the SMM technique. The i-vector fusion between each weight adaptation approach and the classical i-vector yielded slight improvements on the tele- phone part of the NIST 2010 Speaker Recognition Eval- uation dataset.
Copyright (c) 2022, LRE; all rights reserved.
Template by Bootstrapious. Ported to Hugo by DevCows.