default search action
Anton Ragni
Person information
Refine list
refinements active!
zoomed in on ?? of ?? records
view refined list in
export refined list as
2020 – today
- 2024
- [c50]Rhiannon Mogridge, George Close, Robert Sutherland, Thomas Hain, Jon Barker, Stefan Goetze, Anton Ragni:
Non-Intrusive Speech Intelligibility Prediction for Hearing-Impaired Users Using Intermediate ASR Features and Human Memory Models. ICASSP 2024: 306-310 - [c49]Wanli Sun, Zehai Tu, Anton Ragni:
Energy-Based Models for Speech Synthesis. ICASSP 2024: 12667-12671 - [c48]Yizhi Li, Ruibin Yuan, Ge Zhang, Yinghao Ma, Xingran Chen, Hanzhi Yin, Chenghao Xiao, Chenghua Lin, Anton Ragni, Emmanouil Benetos, Norbert Gyenge, Roger B. Dannenberg, Ruibo Liu, Wenhu Chen, Gus Xia, Yemin Shi, Wenhao Huang, Zili Wang, Yike Guo, Jie Fu:
MERT: Acoustic Music Understanding Model with Large-Scale Self-supervised Training. ICLR 2024 - [i20]Rhiannon Mogridge, George Close, Robert Sutherland, Thomas Hain, Jon Barker, Stefan Goetze, Anton Ragni:
Non-Intrusive Speech Intelligibility Prediction for Hearing-Impaired Users using Intermediate ASR Features and Human Memory Models. CoRR abs/2401.13611 (2024) - [i19]Wing-Zin Leung, Mattias Cross, Anton Ragni, Stefan Goetze:
Training Data Augmentation for Dysarthric Automatic Speech Recognition by Text-to-Dysarthric-Speech Synthesis. CoRR abs/2406.08568 (2024) - [i18]Robert Flynn, Anton Ragni:
Self-Train Before You Transcribe. CoRR abs/2406.12937 (2024) - [i17]Yinghao Ma, Anders Øland, Anton Ragni, Bleiz Macsen Del Sette, Charalampos Saitis, Chris Donahue, Chenghua Lin, Christos Plachouras, Emmanouil Benetos, Elio Quinton, Elona Shatri, Fabio Morreale, Ge Zhang, György Fazekas, Gus Xia, Huan Zhang, Ilaria Manco, Jiawen Huang, Julien Guinot, Liwei Lin, Luca Marinelli, Max W. Y. Lam, Megha Sharma, Qiuqiang Kong, Roger B. Dannenberg, Ruibin Yuan, Shangda Wu, Shih-Lun Wu, Shuqi Dai, Shun Lei, Shiyin Kang, Simon Dixon, Wenhu Chen, Wenhao Huang, Xingjian Du, Xingwei Qu, Xu Tan, Yizhi Li, Zeyue Tian, Zhiyong Wu, Zhizheng Wu, Ziyang Ma, Ziyu Wang:
Foundation Models for Music: A Survey. CoRR abs/2408.14340 (2024) - [i16]Mattias Cross, Anton Ragni:
What happens to diffusion model likelihood when your model is conditional? CoRR abs/2409.06364 (2024) - 2023
- [c47]Protima Nomo Sudro, Anton Ragni, Thomas Hain:
Adapting Pretrained Models for Adult to Child Voice Conversion. EUSIPCO 2023: 271-275 - [c46]Robert Flynn, Anton Ragni:
Leveraging Cross-Utterance Context For ASR Decoding. INTERSPEECH 2023: 1359-1363 - [c45]Diane Nicholls, Kate M. Knill, Mark J. F. Gales, Anton Ragni, Paul Ricketts:
Speak & Improve: L2 English Speaking Practice Tool. INTERSPEECH 2023: 3669-3670 - [c44]Yinghao Ma, Ruibin Yuan, Yizhi Li, Ge Zhang, Chenghua Lin, Xingran Chen, Anton Ragni, Hanzhi Yin, Emmanouil Benetos, Norbert Gyenge, Ruibo Liu, Gus Xia, Roger B. Dannenberg, Yike Guo, Jie Fu:
On the Effectiveness of Speech Self-Supervised Learning for Music. ISMIR 2023: 457-465 - [c43]Ruibin Yuan, Yinghao Ma, Yizhi Li, Ge Zhang, Xingran Chen, Hanzhi Yin, Le Zhuo, Yiqi Liu, Jiawen Huang, Zeyue Tian, Binyue Deng, Ningzhi Wang, Chenghua Lin, Emmanouil Benetos, Anton Ragni, Norbert Gyenge, Roger B. Dannenberg, Wenhu Chen, Gus Xia, Wei Xue, Si Liu, Shi Wang, Ruibo Liu, Yike Guo, Jie Fu:
MARBLE: Music Audio Representation Benchmark for Universal Evaluation. NeurIPS 2023 - [i15]Yizhi Li, Ruibin Yuan, Ge Zhang, Yinghao Ma, Xingran Chen, Hanzhi Yin, Chenghua Lin, Anton Ragni, Emmanouil Benetos, Norbert Gyenge, Roger B. Dannenberg, Ruibo Liu, Wenhu Chen, Gus Xia, Yemin Shi, Wenhao Huang, Yike Guo, Jie Fu:
MERT: Acoustic Music Understanding Model with Large-Scale Self-supervised Training. CoRR abs/2306.00107 (2023) - [i14]Ruibin Yuan, Yinghao Ma, Yizhi Li, Ge Zhang, Xingran Chen, Hanzhi Yin, Le Zhuo, Yiqi Liu, Jiawen Huang, Zeyue Tian, Binyue Deng, Ningzhi Wang, Chenghua Lin, Emmanouil Benetos, Anton Ragni, Norbert Gyenge, Roger B. Dannenberg, Wenhu Chen, Gus Xia, Wei Xue, Si Liu, Shi Wang, Ruibo Liu, Yike Guo, Jie Fu:
MARBLE: Music Audio Representation Benchmark for Universal Evaluation. CoRR abs/2306.10548 (2023) - [i13]Robert Flynn, Anton Ragni:
Leveraging Cross-Utterance Context For ASR Decoding. CoRR abs/2306.16903 (2023) - [i12]Yinghao Ma, Ruibin Yuan, Yizhi Li, Ge Zhang, Xingran Chen, Hanzhi Yin, Chenghua Lin, Emmanouil Benetos, Anton Ragni, Norbert Gyenge, Ruibo Liu, Gus Xia, Roger B. Dannenberg, Yike Guo, Jie Fu:
On the Effectiveness of Speech Self-supervised Learning for Music. CoRR abs/2307.05161 (2023) - [i11]Wanli Sun, Zehai Tu, Anton Ragni:
Energy-Based Models For Speech Synthesis. CoRR abs/2310.12765 (2023) - [i10]Robert Flynn, Anton Ragni:
How Much Context Does My Attention-Based ASR System Need? CoRR abs/2310.15672 (2023) - 2022
- [j4]Anton Ragni, Mark J. F. Gales, Oliver Rose, Katherine M. Knill, Alexandros Kastanos, Qiujia Li, Preben Ness:
Increasing Context for Estimating Confidence Scores in Automatic Speech Recognition. IEEE ACM Trans. Audio Speech Lang. Process. 30: 1319-1329 (2022) - [c42]Yizhi Li, Ge Zhang, Bohao Yang, Chenghua Lin, Anton Ragni, Shi Wang, Jie Fu:
HERB: Measuring Hierarchical Regional Bias in Pre-trained Language Models. AACL/IJCNLP (Findings) 2022: 334-346 - [i9]Yizhi Li, Ge Zhang, Bohao Yang, Chenghua Lin, Shi Wang, Anton Ragni, Jie Fu:
HERB: Measuring Hierarchical Regional Bias in Pre-trained Language Models. CoRR abs/2211.02882 (2022) - [i8]Yizhi Li, Ruibin Yuan, Ge Zhang, Yinghao Ma, Chenghua Lin, Xingran Chen, Anton Ragni, Hanzhi Yin, Zhijie Hu, Haoyu He, Emmanouil Benetos, Norbert Gyenge, Ruibo Liu, Jie Fu:
MAP-Music2Vec: A Simple and Effective Baseline for Self-Supervised Music Audio Representation Learning. CoRR abs/2212.02508 (2022) - 2021
- [i7]Sindre André Jacobsen, Anton Ragni:
Continuous representations of intents for dialogue systems. CoRR abs/2105.03716 (2021) - [i6]Zhengxiong Wang, Anton Ragni:
Approximate Fixed-Points in Recurrent Neural Networks. CoRR abs/2106.02417 (2021) - 2020
- [c41]Alexandros Kastanos, Anton Ragni, Mark J. F. Gales:
Confidence Estimation for Black Box Automatic Speech Recognition Systems Using Lattice Recurrent Neural Networks. ICASSP 2020: 6329-6333
2010 – 2019
- 2019
- [j3]Xie Chen, Xunying Liu, Yu Wang, Anton Ragni, Jeremy Heng Meng Wong, Mark J. F. Gales:
Exploiting Future Word Contexts in Neural Network Language Models for Speech Recognition. IEEE ACM Trans. Audio Speech Lang. Process. 27(9): 1444-1454 (2019) - [c40]Suraj Nair, Anton Ragni, Ondrej Klejch, Petra Galuscáková, Douglas W. Oard:
Experiments with Cross-Language Speech Retrieval for Lower-Resource Languages. AIRS 2019: 145-157 - [c39]Douglas W. Oard, Marine Carpuat, Petra Galuscáková, Joseph Barrow, Suraj Nair, Xing Niu, Han-Chin Shing, Weijia Xu, Elena Zotkina, Kathleen R. McKeown, Smaranda Muresan, Efsun Selin Kayi, Ramy Eskander, Chris Kedzie, Yan Virin, Dragomir R. Radev, Rui Zhang, Mark J. F. Gales, Anton Ragni, Kenneth Heafield:
Surprise Languages: Rapid-Response Cross-Language IR. EVIA@NTCIR 2019 - [c38]Qiujia Li, Preben Ness, Anton Ragni, Mark J. F. Gales:
Bi-directional Lattice Recurrent Neural Networks for Confidence Estimation. ICASSP 2019: 6755-6759 - [i5]Alexandros Kastanos, Anton Ragni, Mark J. F. Gales:
Confidence Estimation for Black Box Automatic Speech Recognition Systems Using Lattice Recurrent Neural Networks. CoRR abs/1910.11933 (2019) - 2018
- [j2]Chunyang Wu, Mark J. F. Gales, Anton Ragni, Penny Karanasou, Khe Chai Sim:
Improving Interpretability and Regularization in Deep Learning. IEEE ACM Trans. Audio Speech Lang. Process. 26(2): 256-265 (2018) - [c37]Yu Wang, Xie Chen, Mark J. F. Gales, Anton Ragni, Jeremy Heng Meng Wong:
Phonetic and Graphemic Systems for Multi-Genre Broadcast Transcription. ICASSP 2018: 5899-5903 - [c36]Kate M. Knill, Mark J. F. Gales, Konstantinos Kyriakopoulos, Andrey Malinin, Anton Ragni, Yu Wang, Andrew Caines:
Impact of ASR Performance on Free Speaking Language Assessment. INTERSPEECH 2018: 1641-1645 - [c35]Anton Ragni, Mark J. F. Gales:
Automatic Speech Recognition System Development in the "Wild". INTERSPEECH 2018: 2217-2221 - [c34]Oscar Chen, Anton Ragni, Mark J. F. Gales, Xie Chen:
Active Memory Networks for Language Modeling. INTERSPEECH 2018: 3338-3342 - [c33]Anton Ragni, Qiujia Li, Mark J. F. Gales, Yongqiang Wang:
Confidence Estimation and Deletion Prediction Using Bidirectional Recurrent Neural Networks. SLT 2018: 204-211 - [c32]Yu Wang, Jeremy Heng Meng Wong, Mark J. F. Gales, Kate M. Knill, Anton Ragni:
Sequence Teacher-Student Training of Acoustic Models for Automatic Free Speaking Language Assessment. SLT 2018: 994-1000 - [i4]Yu Wang, Xie Chen, Mark J. F. Gales, Anton Ragni, Jeremy Heng Meng Wong:
Phonetic and Graphemic Systems for Multi-Genre Broadcast Transcription. CoRR abs/1802.00254 (2018) - [i3]Qiujia Li, Preben Ness, Anton Ragni, Mark J. F. Gales:
Bi-Directional Lattice Recurrent Neural Networks for Confidence Estimation. CoRR abs/1810.13024 (2018) - [i2]Anton Ragni, Qiujia Li, Mark J. F. Gales, Yu Wang:
Confidence Estimation and Deletion Prediction Using Bidirectional Recurrent Neural Networks. CoRR abs/1810.13025 (2018) - 2017
- [c31]Andrey Malinin, Anton Ragni, Kate M. Knill, Mark J. F. Gales:
Incorporating Uncertainty into Deep Learning for Spoken Language Assessment. ACL (2) 2017: 45-50 - [c30]Xie Chen, X. Liu, Anton Ragni, Y. Wang, Mark J. F. Gales:
Future word contexts in neural network language models. ASRU 2017: 97-103 - [c29]Anton Ragni, Chunyang Wu, Mark J. F. Gales, J. Vasilakes, Kate M. Knill:
Stimulated training for automatic speech recognition and keyword search in limited resource conditions. ICASSP 2017: 4830-4834 - [c28]Anton Ragni, Danielle Saunders, P. Zahemszky, J. Vasilakes, Mark J. F. Gales, Kate M. Knill:
Morph-to-word transduction for accurate and efficient automatic speech recognition and keyword search. ICASSP 2017: 5770-5774 - [c27]Xie Chen, Anton Ragni, J. Vasilakes, Xunying Liu, Kate M. Knill, Mark J. F. Gales:
Recurrent neural network language models for keyword search. ICASSP 2017: 5775-5779 - [c26]Xie Chen, Anton Ragni, Xunying Liu, Mark J. F. Gales:
Investigating Bidirectional Recurrent Neural Network Language Models for Speech Recognition. INTERSPEECH 2017: 269-273 - [c25]Kate M. Knill, Mark J. F. Gales, Konstantinos Kyriakopoulos, Anton Ragni, Yu Wang:
Use of Graphemic Lexicons for Spoken Language Assessment. INTERSPEECH 2017: 2774-2778 - [c24]Andrey Malinin, Kate M. Knill, Anton Ragni, Yu Wang, Mark J. F. Gales:
An attention based model for off-topic spontaneous spoken response detection: An Initial Study. SLaTE 2017: 144-149 - [c23]Mark J. F. Gales, Kate M. Knill, Anton Ragni:
Low-Resource Speech Recognition and Keyword-Spotting. SPECOM 2017: 3-19 - [i1]Xie Chen, Xunying Liu, Anton Ragni, Yu Wang, Mark J. F. Gales:
Future Word Contexts in Neural Network Language Models. CoRR abs/1708.05592 (2017) - 2016
- [c22]J. Yang, Chao Zhang, Anton Ragni, Mark J. F. Gales, Philip C. Woodland:
System combination with log-linear models. ICASSP 2016: 5675-5679 - [c21]Jingzhou Yang, Anton Ragni, Mark J. F. Gales, Kate M. Knill:
Log-Linear System Combination Using Structured Support Vector Machines. INTERSPEECH 2016: 1898-1902 - [c20]Anton Ragni, Edgar Dakin, Xie Chen, Mark J. F. Gales, Kate M. Knill:
Multi-Language Neural Network Language Models. INTERSPEECH 2016: 3042-3046 - 2015
- [c19]Rogier C. van Dalen, Jingzhou Yang, Haipeng Wang, Anton Ragni, Chao Zhang, Mark J. F. Gales:
Structured discriminative models using deep neural-network features. ASRU 2015: 160-166 - [c18]Jia Cui, Brian Kingsbury, Bhuvana Ramabhadran, Abhinav Sethy, Kartik Audhkhasi, Xiaodong Cui, Ellen Kislal, Lidia Mangu, Markus Nußbaum-Thom, Michael Picheny, Zoltán Tüske, Pavel Golik, Ralf Schlüter, Hermann Ney, Mark J. F. Gales, Kate M. Knill, Anton Ragni, Haipeng Wang, Philip C. Woodland:
Multilingual representations for low resource speech recognition and keyword search. ASRU 2015: 259-266 - [c17]Anton Ragni, Mark J. F. Gales, Kate M. Knill:
A language space representation for speech recognition. ICASSP 2015: 4634-4638 - [c16]Mark J. F. Gales, Kate M. Knill, Anton Ragni:
Unicode-based graphemic systems for limited resource languages. ICASSP 2015: 5186-5190 - [c15]Gideon Mendels, Erica Cooper, Victor Soto, Julia Hirschberg, Mark J. F. Gales, Kate M. Knill, Anton Ragni, Haipeng Wang:
Improving speech recognition and keyword search for low resource languages using web data. INTERSPEECH 2015: 829-833 - [c14]Haipeng Wang, Anton Ragni, Mark J. F. Gales, Kate M. Knill, Philip C. Woodland, Chao Zhang:
Joint decoding of tandem and hybrid systems for improved keyword spotting on low resource languages. INTERSPEECH 2015: 3660-3664 - 2014
- [c13]Takuya Yoshioka, Anton Ragni, Mark J. F. Gales:
Investigation of unsupervised adaptation of DNN acoustic models with filter bank input. ICASSP 2014: 6344-6348 - [c12]Kate M. Knill, Mark J. F. Gales, Anton Ragni, Shakti P. Rath:
Language independent and unsupervised acoustic models for speech recognition and keyword spotting. INTERSPEECH 2014: 16-20 - [c11]Anton Ragni, Kate M. Knill, Shakti P. Rath, Mark J. F. Gales:
Data augmentation for low resource languages. INTERSPEECH 2014: 810-814 - [c10]Shakti P. Rath, Kate M. Knill, Anton Ragni, Mark J. F. Gales:
Combining tandem and hybrid systems for improved speech recognition and keyword spotting on low resource languages. INTERSPEECH 2014: 835-839 - [c9]Mark J. F. Gales, Kate M. Knill, Anton Ragni, Shakti P. Rath:
Speech recognition and keyword spotting for low-resource languages: Babel project research at CUED. SLTU 2014: 16-23 - 2013
- [c8]Rogier C. van Dalen, Anton Ragni, Mark J. F. Gales:
Efficient decoding with generative score-spaces using the expectation semiring. ICASSP 2013: 7619-7623 - 2012
- [c7]Anton Ragni, Mark J. F. Gales:
Inference algorithms for generative score-spaces. ICASSP 2012: 4149-4152 - [c6]Zoi Roupakia, Anton Ragni, Mark J. F. Gales:
Rapid Nonlinear Speaker Adaptation for Large-Vocabulary Continuous Speech Recognition. INTERSPEECH 2012: 1784-1787 - [c5]Mark J. F. Gales, Anton Ragni, Austin Zhang, Rogier C. van Dalen:
Structured discriminative models for speech recognition. MLSLP 2012 - 2011
- [c4]Anton Ragni, Mark J. F. Gales:
Derivative kernels for noise robust ASR. ASRU 2011: 119-124 - [c3]Anton Ragni, Mark John Francis Gales:
Structured discriminative models for noise robust continuous speech recognition. ICASSP 2011: 4788-4791 - 2010
- [j1]Shi-Xiong Zhang, Anton Ragni, Mark J. F. Gales:
Structured Log Linear Models for Noise Robust Speech Recognition. IEEE Signal Process. Lett. 17(11): 945-948 (2010)
2000 – 2009
- 2009
- [c2]Mark J. F. Gales, Anton Ragni, H. AlDamarki, C. Gautier:
Support vector machines for noise robust ASR. ASRU 2009: 205-210 - 2007
- [c1]Anton Ragni:
Initial Experiments with Estonian Speech Recognition. NODALIDA 2007: 249-252
Coauthor Index
aka: Mark John Francis Gales
aka: Katherine M. Knill
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.
Unpaywalled article links
Add open access links from to the list of external document links (if available).
Privacy notice: By enabling the option above, your browser will contact the API of unpaywall.org to load hyperlinks to open access articles. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Unpaywall privacy policy.
Archived links via Wayback Machine
For web page which are no longer available, try to retrieve content from the of the Internet Archive (if available).
Privacy notice: By enabling the option above, your browser will contact the API of archive.org to check for archived content of web pages that are no longer available. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Internet Archive privacy policy.
Reference lists
Add a list of references from , , and to record detail pages.
load references from crossref.org and opencitations.net
Privacy notice: By enabling the option above, your browser will contact the APIs of crossref.org, opencitations.net, and semanticscholar.org to load article reference information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Crossref privacy policy and the OpenCitations privacy policy, as well as the AI2 Privacy Policy covering Semantic Scholar.
Citation data
Add a list of citing articles from and to record detail pages.
load citations from opencitations.net
Privacy notice: By enabling the option above, your browser will contact the API of opencitations.net and semanticscholar.org to load citation information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the OpenCitations privacy policy as well as the AI2 Privacy Policy covering Semantic Scholar.
OpenAlex data
Load additional information about publications from .
Privacy notice: By enabling the option above, your browser will contact the API of openalex.org to load additional information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the information given by OpenAlex.
last updated on 2024-10-22 20:17 CEST by the dblp team
all metadata released as open data under CC0 1.0 license
see also: Terms of Use | Privacy Policy | Imprint