The OpenASR (Open Automatic Speech Recognition) 2021 Challenge was the third open challenge and the second OpenASR challenge associated with the IARPA MATERIAL program, after the OpenCLIR (Open Cross-Language Information Retrieval) 2019 Challenge and the OpenASR 2020 Challenge Capabilities tested in these open challenges are expected to ultimately support the MATERIAL task of effective triage and analysis of large volumes of text and audio content in a variety of less-studied languages. OpenASR21 was implemented as a track of NIST’s OpenSAT (Speech Analytic Technologies) evaluation series.
The goal of the OpenASR21 challenge was to assess the state of the art of automatic speech recognition (ASR) technologies for low-resource languages. ASR was performed on speech datasets, and written text output had to be produced.
Please refer to the OpenASR21 Challenge Evaluation Plan for a full description of the challenge and its rules and procedures.
OpenASR21 was offered for 15 low-resource languages, of which participating teams could attempt as many as they wished. Ten of these languages were repeated from OpenASR20 with unchanged datasets; five languages were new. All languages were offered with a case-insensitive scoring dataset. Of the five new languages, three were offered with an additional case-sensitive scoring dataset. The case-sensitive scoring was to serve as a proxy for assessment of ASR performance on proper nouns. The new languages are marked as such below:
The data for the challenge consisted of speech data stemming from the IARPA Babel program, with the exception of Somali. which stemmed from the IARPA MATERIAL program. More details regarding technical data details can be found in section 3 of the IARPA Babel Data Specifications for Performers. For each language (and case scoring condition, where applicable), separate training, development, and evaluation datasets were provided.
For CIS, the datasets for most of the languages stemmed from the IARPA Babel program. The Somali and Farsi datasets stemmed from the IARPA MATERIAL program. For the case-sensitive scoring datasets, all data stemmed from the IARPA MATERIAL program.
The data consisted of conversational speech for the case-insensitive scoring datasets. For the case-sensitive scoring datasets, the data consisted of a mix of three genres: conversational speech, news broadcast, and topical broadcast.
The challenge offered three training conditions. The Constrained-plus training was new for OpenASR21.
The most important milestones of the schedule of the challenge were as follows:
Table 1 lists the teams that submitted valid output, and for which languages they submitted valid output.
26 teams from 13 countries originally registered to participate. Fifteen teams submitted valid output for at least one language under the Constrained training condition.
The table marks the following issues in the pertinent table cells:
Organization | Team | AMH | CAN | FAR | GEO | GUA | JAV | KAZ | KUR | MON | PAS | SOM | SWA | TAG | TAM | VIE |
Shanghai Jian Qiao University, China | Baymax (no system description) | Yes (no Constrained submission) | Yes | |||||||||||||
Catskills Research Company, USA | Catskills | Yes | Yes | |||||||||||||
Deep Learning and Media System Laboratory, National Central University, Taiwan, China | CHNGA-DLMSL (no system description) | Yes | Yes (late) | |||||||||||||
Centre de Recherche Informatique de Montréal, Canada | CRIM | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes |
Northeastern University, China | FiberHome-AILab (no system description) | Yes | ||||||||||||||
Institute for Infocomm Research, Singapore | I2R | Yes | Yes | Yes (late) | Yes | |||||||||||
University of Exeter, UK | Jarvs (no system description) | Yes (no Constrained submission) | ||||||||||||||
Tencent, China | MMT | Yes (no Constrained submission) | Yes (no Constrained submission) | Yes | ||||||||||||
Samsung R&D Institute Bangalore, India | SRIB_ASR | Yes | ||||||||||||||
Tallinn University of Technology, Estonia | TalTech | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes |
Tsinghua University, China | THUEE | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes |
Tencent & Tsinghua University, China | TNT | Yes | Yes (no Constrained submission) | Yes | ||||||||||||
Transsion AI Technology Dept., China | TranSpeech (no system description) | Yes (no Constrained submission) | Yes (late, no Constrained submission) | Yes | ||||||||||||
National Engineering Laboratory for Speech and Language Information Processing, University of Science and Technology of China | ustc_nelslip | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes |
SpeakIn Technologies Co., Ltd, China | YCG-AI (no system description) | Yes | ||||||||||||||
University of Science and Technology of China | zxy | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes | Yes |
Table 1: OpenASR21 Participating teams and languages.
Table 2 lists the best WER result achieved by each team, ordered by language, scoring, training condition, and WER score. The CER score for the same submission is also provided.
The table marks the following issues in the Notes column:
Self-reported time and memory resources are not included in this overview of results.
On-time submissions | ||||||
Language | Scoring | Training | Team | WER | CER | Notes |
Amharic | Case-insensitive | Constrained | ustc_nelslip | 0.3993 | 0.3019 | |
Amharic | Case-insensitive | Constrained | THUEE | 0.4219 | 0.3331 | |
Amharic | Case-insensitive | Constrained | zxy | 0.4257 | 0.3114 | |
Amharic | Case-insensitive | Constrained | CRIM | 0.4310 | 0.3288 | |
Amharic | Case-insensitive | Constrained | TalTech | 0.4318 | 0.3277 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Amharic | Case-insensitive | Constrained-plus | THUEE | 0.3729 | 0.2753 | |
Amharic | Case-insensitive | Constrained-plus | zxy | 0.4045 | 0.2986 | |
Amharic | Case-insensitive | Constrained-plus | TalTech | 0.4152 | 0.3101 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Amharic | Case-insensitive | Unconstrained | TalTech | 0.3522 | 0.2609 | |
Amharic | Case-insensitive | Unconstrained | TranSpeech | 0.4424 | 0.3167 | No Constrained submission, no system description |
Amharic | Case-insensitive | Unconstrained | Jarvs | 0.4446 | 0.3191 | No Constrained submission, no system description |
Amharic | Case-insensitive | Unconstrained | Baymax | 0.5100 | 0.3044 | No Constrained submission, no system description |
Language | Scoring | Training | Team | WER | CER | Notes |
Cantonese | Case-insensitive | Constrained | ustc_nelslip | 0.3762 | 0.3274 | |
Cantonese | Case-insensitive | Constrained | TNT | 0.4024 | 0.3511 | |
Cantonese | Case-insensitive | Constrained | THUEE | 0.4038 | 0.3559 | |
Cantonese | Case-insensitive | Constrained | TalTech | 0.4267 | 0.3774 | |
Cantonese | Case-insensitive | Constrained | CRIM | 0.4277 | 0.3717 | |
Cantonese | Case-insensitive | Constrained | zxy | 0.4830 | 0.4245 | |
Cantonese | Case-insensitive | Constrained | CHNGA-DLMSL | 0.7113 | 0.6571 | No system description |
Cantonese | Case-insensitive | Constrained | YCG-AI | 0.7547 | 0.6669 | No system description |
Cantonese | Case-insensitive | Constrained | FiberHome-AILab | 1.2414 | 1.2284 | No system description |
Language | Scoring | Training | Team | WER | CER | Notes |
Cantonese | Case-insensitive | Constrained-plus | TNT | 0.3371 | 0.2847 | |
Cantonese | Case-insensitive | Constrained-plus | MMT | 0.3631 | 0.3011 | No Constrained submission |
Cantonese | Case-insensitive | Constrained-plus | zxy | 0.3880 | 0.3337 | |
Cantonese | Case-insensitive | Constrained-plus | TalTech | 0.4101 | 0.3608 | |
Cantonese | Case-insensitive | Constrained-plus | CHNGA-DLMSL | 0.6834 | 0.6003 | No system description |
Cantonese | Case-insensitive | Constrained-plus | YCG-AI | 0.7091 | 0.5903 | No system description |
Language | Scoring | Training | Team | WER | CER | Notes |
Cantonese | Case-insensitive | Unconstrained | ustc_nelslip | 0.2665 | 0.1982 | |
Cantonese | Case-insensitive | Unconstrained | TNT | 0.2776 | 0.2239 | |
Cantonese | Case-insensitive | Unconstrained | MMT | 0.2864 | 0.2332 | No Constrained submission |
Cantonese | Case-insensitive | Unconstrained | zxy | 0.3019 | 0.2238 | |
Cantonese | Case-insensitive | Unconstrained | YCG-AI | 0.7710 | 0.6246 | No system description |
Language | Scoring | Training | Team | WER | CER | Notes |
Farsi | Case-insensitive | Constrained | ustc_nelslip | 0.6796 | 0.5472 | |
Farsi | Case-insensitive | Constrained | zxy | 0.6986 | 0.5366 | |
Farsi | Case-insensitive | Constrained | CRIM | 0.7932 | 0.6660 | |
Farsi | Case-insensitive | Constrained | THUEE | 0.8144 | 0.7533 | |
Farsi | Case-insensitive | Constrained | TalTech | 0.8159 | 0.7533 | |
Farsi | Case-insensitive | Constrained | Catskills | 0.9865 | 0.8898 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Farsi | Case-insensitive | Constrained-plus | THUEE | 0.6262 | 0.4978 | |
Farsi | Case-insensitive | Constrained-plus | zxy | 0.6980 | 0.5800 | |
Farsi | Case-insensitive | Constrained-plus | TalTech | 0.7906 | 0.7261 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Farsi | Case-insensitive | Unconstrained | ustc_nelslip | 0.5200 | 0.3691 | |
Farsi | Case-insensitive | Unconstrained | zxy | 0.5200 | 0.3691 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Georgian | Case-insensitive | Constrained | ustc_nelslip | 0.3918 | 0.2526 | |
Georgian | Case-insensitive | Constrained | TalTech | 0.4155 | 0.2632 | |
Georgian | Case-insensitive | Constrained | zxy | 0.4231 | 0.2665 | |
Georgian | Case-insensitive | Constrained | CRIM | 0.4255 | 0.2691 | |
Georgian | Case-insensitive | Constrained | THUEE | 0.4657 | 0.3296 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Georgian | Case-insensitive | Constrained-plus | TalTech | 0.3894 | 0.2415 | |
Georgian | Case-insensitive | Constrained-plus | zxy | 0.3978 | 0.2504 | |
Georgian | Case-insensitive | Constrained-plus | THUEE | 0.4039 | 0.2552 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Guarani | Case-insensitive | Constrained | ustc_nelslip | 0.4262 | 0.3939 | |
Guarani | Case-insensitive | Constrained | THUEE | 0.4423 | 0.4099 | |
Guarani | Case-insensitive | Constrained | TalTech | 0.4524 | 0.4168 | |
Guarani | Case-insensitive | Constrained | zxy | 0.4548 | 0.4208 | |
Guarani | Case-insensitive | Constrained | CRIM | 0.4597 | 0.4297 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Guarani | Case-insensitive | Constrained-plus | THUEE | 0.4058 | 0.3706 | |
Guarani | Case-insensitive | Constrained-plus | zxy | 0.4312 | 0.3954 | |
Guarani | Case-insensitive | Constrained-plus | TalTech | 0.4375 | 0.4034 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Javanese | Case-insensitive | Constrained | ustc_nelslip | 0.4808 | 0.4813 | |
Javanese | Case-insensitive | Constrained | zxy | 0.5110 | 0.5115 | |
Javanese | Case-insensitive | Constrained | CRIM | 0.5197 | 0.5201 | |
Javanese | Case-insensitive | Constrained | THUEE | 0.5210 | 0.5216 | |
Javanese | Case-insensitive | Constrained | TalTech | 0.5272 | 0.5276 | |
Javanese | Case-insensitive | Constrained | I2R | 0.8957 | 0.8961 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Javanese | Case-insensitive | Constrained-plus | THUEE | 0.4797 | 0.4801 | |
Javanese | Case-insensitive | Constrained-plus | zxy | 0.4838 | 0.4845 | |
Javanese | Case-insensitive | Constrained-plus | TalTech | 0.5058 | 0.5063 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Javanese | Case-insensitive | Unconstrained | ustc_nelslip | 0.3951 | 0.3920 | |
Javanese | Case-insensitive | Unconstrained | zxy | 0.4442 | 0.4450 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Kazakh | Case-insensitive | Constrained | ustc_nelslip | 0.4997 | 0.3794 | |
Kazakh | Case-insensitive | Constrained | TalTech | 0.5300 | 0.3817 | |
Kazakh | Case-insensitive | Constrained | zxy | 0.5334 | 0.3696 | |
Kazakh | Case-insensitive | Constrained | THUEE | 0.5454 | 0.4373 | |
Kazakh | Case-insensitive | Constrained | CRIM | 0.5685 | 0.4378 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Kazakh | Case-insensitive | Constrained-plus | TNT | 0.4279 | 0.2903 | No Constrained submission |
Kazakh | Case-insensitive | Constrained-plus | THUEE | 0.4287 | 0.2889 | |
Kazakh | Case-insensitive | Constrained-plus | zxy | 0.4991 | 0.3606 | |
Kazakh | Case-insensitive | Constrained-plus | TalTech | 0.5058 | 0.3646 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Kazakh | Case-insensitive | Unconstrained | ustc_nelslip | 0.3745 | 0.2499 | |
Kazakh | Case-insensitive | Unconstrained | zxy | 0.3812 | 0.2545 | |
Kazakh | Case-insensitive | Unconstrained | TNT | 0.3945 | 0.2928 | No Constrained submission |
Kazakh | Case-insensitive | Unconstrained | MMT | 0.4036 | 0.2928 | No Constrained submission |
Language | Scoring | Training | Team | WER | CER | Notes |
Kazakh | Case-sensitive | Constrained | THUEE | 0.4983 | 0.2900 | |
Kazakh | Case-sensitive | Constrained | zxy | 0.5468 | 0.2931 | |
Kazakh | Case-sensitive | Constrained | CRIM | 0.5855 | 0.3206 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Kazakh | Case-sensitive | Constrained-plus | THUEE | 0.4989 | 0.3282 | |
Kazakh | Case-sensitive | Constrained-plus | zxy | 0.5285 | 0.3026 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Kurmanji-Kurdish | Case-insensitive | Constrained | ustc_nelslip | 0.6168 | 0.5776 | |
Kurmanji-Kurdish | Case-insensitive | Constrained | TalTech | 0.6437 | 0.6079 | |
Kurmanji-Kurdish | Case-insensitive | Constrained | zxy | 0.6537 | 0.6076 | |
Kurmanji-Kurdish | Case-insensitive | Constrained | CRIM | 0.6567 | 0.6173 | |
Kurmanji-Kurdish | Case-insensitive | Constrained | THUEE | 0.6585 | 0.6272 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Kurmanji-Kurdish | Case-insensitive | Constrained-plus | THUEE | 0.6154 | 0.5673 | |
Kurmanji-Kurdish | Case-insensitive | Constrained-plus | zxy | 0.6185 | 0.5809 | |
Kurmanji-Kurdish | Case-insensitive | Constrained-plus | TalTech | 0.6213 | 0.5846 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Mongolian | Case-insensitive | Constrained | ustc_nelslip | 0.4098 | 0.3025 | |
Mongolian | Case-insensitive | Constrained | MMT | 0.4165 | 0.3048 | |
Mongolian | Case-insensitive | Constrained | TNT | 0.4171 | 0.3006 | |
Mongolian | Case-insensitive | Constrained | zxy | 0.4489 | 0.3169 | |
Mongolian | Case-insensitive | Constrained | TalTech | 0.4531 | 0.3343 | |
Mongolian | Case-insensitive | Constrained | THUEE | 0.4531 | 0.3530 | |
Mongolian | Case-insensitive | Constrained | CRIM | 0.4603 | 0.3357 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Mongolian | Case-insensitive | Constrained-plus | TNT | 0.3781 | 0.2824 | |
Mongolian | Case-insensitive | Constrained-plus | MMT | 0.3787 | 0.2825 | |
Mongolian | Case-insensitive | Constrained-plus | THUEE | 0.4113 | 0.3002 | |
Mongolian | Case-insensitive | Constrained-plus | zxy | 0.4190 | 0.3047 | |
Mongolian | Case-insensitive | Constrained-plus | TalTech | 0.4366 | 0.3211 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Mongolian | Case-insensitive | Unconstrained | ustc_nelslip | 0.3152 | 0.2509 | |
Mongolian | Case-insensitive | Unconstrained | zxy | 0.3188 | 0.2514 | |
Mongolian | Case-insensitive | Unconstrained | TNT | 0.3430 | 0.2525 | |
Mongolian | Case-insensitive | Unconstrained | MMT | 0.3728 | 0.2781 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Pashto | Case-insensitive | Constrained | ustc_nelslip | 0.4322 | 0.3045 | |
Pashto | Case-insensitive | Constrained | TalTech | 0.4530 | 0.3236 | |
Pashto | Case-insensitive | Constrained | THUEE | 0.4639 | 0.3302 | |
Pashto | Case-insensitive | Constrained | CRIM | 0.4723 | 0.3256 | |
Pashto | Case-insensitive | Constrained | zxy | 0.4734 | 0.3122 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Pashto | Case-insensitive | Constrained-plus | THUEE | 0.4138 | 0.2812 | |
Pashto | Case-insensitive | Constrained-plus | TalTech | 0.4249 | 0.2950 | |
Pashto | Case-insensitive | Constrained-plus | zxy | 0.4401 | 0.2993 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Pashto | Case-insensitive | Unconstrained | ustc_nelslip | 0.3388 | 0.2288 | |
Pashto | Case-insensitive | Unconstrained | zxy | 0.3403 | 0.2301 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Somali | Case-insensitive | Constrained | ustc_nelslip | 0.5560 | 0.5570 | |
Somali | Case-insensitive | Constrained | THUEE | 0.5802 | 0.5810 | |
Somali | Case-insensitive | Constrained | zxy | 0.5855 | 0.5866 | |
Somali | Case-insensitive | Constrained | TalTech | 0.5871 | 0.5879 | |
Somali | Case-insensitive | Constrained | CRIM | 0.5923 | 0.5933 | |
Somali | Case-insensitive | Constrained | Catskills | 0.9931 | 0.9931 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Somali | Case-insensitive | Constrained-plus | THUEE | 0.5334 | 0.5345 | |
Somali | Case-insensitive | Constrained-plus | zxy | 0.5522 | 0.5534 | |
Somali | Case-insensitive | Constrained-plus | TalTech | 0.5690 | 0.5700 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Swahili | Case-insensitive | Constrained | ustc_nelslip | 0.3242 | 0.3259 | |
Swahili | Case-insensitive | Constrained | zxy | 0.3473 | 0.3489 | |
Swahili | Case-insensitive | Constrained | TalTech | 0.3498 | 0.3512 | |
Swahili | Case-insensitive | Constrained | CRIM | 0.3501 | 0.3513 | |
Swahili | Case-insensitive | Constrained | THUEE | 0.3524 | 0.3542 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Swahili | Case-insensitive | Constrained-plus | THUEE | 0.3149 | 0.3168 | |
Swahili | Case-insensitive | Constrained-plus | zxy | 0.3285 | 0.3307 | |
Swahili | Case-insensitive | Constrained-plus | TalTech | 0.3309 | 0.3324 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Swahili | Case-sensitive | Constrained | zxy | 0.4350 | 0.4353 | |
Swahili | Case-sensitive | Constrained | THUEE | 0.4371 | 0.4375 | |
Swahili | Case-sensitive | Constrained | CRIM | 0.4853 | 0.4857 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Swahili | Case-sensitive | Constrained-plus | THUEE | 0.4394 | 0.4399 | |
Swahili | Case-sensitive | Constrained-plus | zxy | 0.4399 | 0.4402 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Tagalog | Case-insensitive | Constrained | ustc_nelslip | 0.4038 | 0.4042 | |
Tagalog | Case-insensitive | Constrained | THUEE | 0.4170 | 0.4174 | |
Tagalog | Case-insensitive | Constrained | TalTech | 0.4186 | 0.4194 | |
Tagalog | Case-insensitive | Constrained | CRIM | 0.4318 | 0.4325 | |
Tagalog | Case-insensitive | Constrained | zxy | 0.4369 | 0.4375 | |
Tagalog | Case-insensitive | Constrained | TranSpeech | 0.7195 | 0.7200 | No system description |
Tagalog | Case-insensitive | Constrained | Baymax | 0.7445 | 0.7450 | No system description |
Tagalog | Case-insensitive | Constrained | I2R | 0.8366 | 0.8366 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Tagalog | Case-insensitive | Constrained-plus | THUEE | 0.3767 | 0.3772 | |
Tagalog | Case-insensitive | Constrained-plus | TalTech | 0.4017 | 0.4024 | |
Tagalog | Case-insensitive | Constrained-plus | zxy | 0.4132 | 0.4140 | |
Tagalog | Case-insensitive | Constrained-plus | Baymax | 0.7961 | 0.7963 | No system description |
Language | Scoring | Training | Team | WER | CER | Notes |
Tagalog | Case-sensitive | Constrained | zxy | 0.4618 | 0.4617 | |
Tagalog | Case-sensitive | Constrained | THUEE | 0.4900 | 0.4896 | |
Tagalog | Case-sensitive | Constrained | CRIM | 0.5323 | 0.5322 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Tagalog | Case-sensitive | Constrained-plus | zxy | 0.4630 | 0.4629 | |
Tagalog | Case-sensitive | Constrained-plus | THUEE | 0.4641 | 0.4648 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Tamil | Case-insensitive | Constrained | ustc_nelslip | 0.6226 | 0.4130 | |
Tamil | Case-insensitive | Constrained | CRIM | 0.6382 | 0.4107 | |
Tamil | Case-insensitive | Constrained | THUEE | 0.6408 | 0.4482 | |
Tamil | Case-insensitive | Constrained | TalTech | 0.6423 | 0.4190 | |
Tamil | Case-insensitive | Constrained | zxy | 0.6577 | 0.4215 | |
Tamil | Case-insensitive | Constrained | SRIB_ASR | 0.8253 | 0.5787 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Tamil | Case-insensitive | Constrained-plus | THUEE | 0.5962 | 0.3810 | |
Tamil | Case-insensitive | Constrained-plus | TalTech | 0.6283 | 0.4055 | |
Tamil | Case-insensitive | Constrained-plus | zxy | 0.6323 | 0.4031 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Tamil | Case-insensitive | Unconstrained | ustc_nelslip | 0.5593 | 0.3859 | |
Tamil | Case-insensitive | Unconstrained | zxy | 0.5695 | 0.3966 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Vietnamese | Case-insensitive | Constrained | ustc_nelslip | 0.4031 | 0.3605 | |
Vietnamese | Case-insensitive | Constrained | TalTech | 0.4310 | 0.3899 | |
Vietnamese | Case-insensitive | Constrained | zxy | 0.4344 | 0.3879 | |
Vietnamese | Case-insensitive | Constrained | CRIM | 0.4398 | 0.3981 | |
Vietnamese | Case-insensitive | Constrained | THUEE | 0.4445 | 0.4016 | |
Vietnamese | Case-insensitive | Constrained | I2R | 0.7981 | 0.7404 | |
Language | Scoring | Training | Team | WER | CER | Notes |
Vietnamese | Case-insensitive | Constrained-plus | THUEE | 0.3727 | 0.3127 | |
Vietnamese | Case-insensitive | Constrained-plus | TalTech | 0.4010 | 0.3610 | |
Vietnamese | Case-insensitive | Constrained-plus | zxy | 0.4044 | 0.3609 | |
Late submissions | ||||||
Language | Scoring | Training | Team | WER | CER | Notes |
Cantonese | Case-insensitive | Constrained | CHNGA-DLMSL | 0.6894 | 0.6331 | Late, no system description |
Language | Scoring | Training | Team | WER | CER | Notes |
Cantonese | Case-insensitive | Constrained-plus | CHNGA-DLMSL | 0.6696 | 0.5848 | Late, no system description |
Language | Scoring | Training | Team | WER | CER | Notes |
Farsi | Case-insensitive | Constrained | CHNGA-DLMSL | 0.9626 | 0.6339 | Late, no system description |
Language | Scoring | Training | Team | WER | CER | Notes |
Kazakh | Case-sensitive | Constrained | zxy | 0.5106 | 0.3092 | Late |
Language | Scoring | Training | Team | WER | CER | Notes |
Kurmanji-Kurdish | Case-insensitive | Constrained-plus | THUEE | 0.6019 | 0.5680 | Late |
Language | Scoring | Training | Team | WER | CER | Notes |
Swahili | Case-insensitive | Unconstrained | TranSpeech | 0.4218 | 0.4239 | Late, no Constrained submission, no system description |
Language | Scoring | Training | Team | WER | CER | Notes |
Tamil | Case-insensitive | Constrained | THUEE | 0.6391 | 0.4534 | Late |
Tamil | Case-insensitive | Constrained | SRIB_ASR | 0.7942 | 0.5772 | Late |
Tamil | Case-insensitive | Constrained | I2R | 1.0170 | 0.8558 | Late |
Language | Scoring | Training | Team | WER | CER | Notes |
Vietnamese | Case-insensitive | Constrained-plus | THUEE | 0.3574 | 0.3148 | Late |
Table 2: OpenASR21 Results. Submissions with best WER for each participating team by language, scoring condition, and training condition.
As part of the evaluation submission, teams were required to include a paper to describe their systems. Submitted system descriptions are provided below.
Teams were also encouraged to submit their updated work to be included in a Low-Resource ASR Development Special Session at INTERSPEECH 2022.
NIST serves to coordinate the evaluations in order to support research and to help advance the state- of-the-art. NIST evaluations are not viewed as a competition, and such results reported by NIST are not to be construed, or represented, as endorsements of any team's system, or as official findings on the part of NIST or the U.S. Government.
Please email openasr_poc [at] nist.gov (openasr_poc[at]nist[dot]gov) for any questions or comments regarding the OpenASR Challenge.