Abstract
Automatic speech recognition has gone through many changes in recent years. Advances both in computer hardware and machine learning have made it possible to develop systems far more capable and complex than the previous state-of-the-art. However, almost all of these improvements have been tested in major well-resourced languages.
In this paper, we show that these techniques are capable of yielding improvements even in a small data scenario. We experiment with different deep neural network architectures for acoustic modeling for Northern Sámi, and report up to 50% relative error rate reductions.
We also run experiments to compare the performance of different subwords as language modeling units in Northern Sámi.
In this paper, we show that these techniques are capable of yielding improvements even in a small data scenario. We experiment with different deep neural network architectures for acoustic modeling for Northern Sámi, and report up to 50% relative error rate reductions.
We also run experiments to compare the performance of different subwords as language modeling units in Northern Sámi.
Original language | English |
---|---|
Title of host publication | Fourth International Workshop on Computational Linguistics for Uralic Languages |
Publisher | Association for Computational Linguistics |
Pages | 89-99 |
Number of pages | 11 |
DOIs | |
Publication status | Published - 2017 |
MoE publication type | D3 Professional conference proceedings |
Event | International Workshop on Computational Linguistics for the Uralic Languages - Helsinki, Finland Duration: 8 Jan 2018 → 9 Jan 2018 Conference number: 4 |
Workshop
Workshop | International Workshop on Computational Linguistics for the Uralic Languages |
---|---|
Abbreviated title | IWCLUL |
Country/Territory | Finland |
City | Helsinki |
Period | 08/01/2018 → 09/01/2018 |