Multi-Task and Transfer Learning in Low-Resource Speech Recognition
dc.contributor.advisor | Bever, Thomas G. | |
dc.contributor.advisor | Surdeanu, Mihai | |
dc.contributor.author | Meyer, Josh | |
dc.creator | Meyer, Josh | |
dc.date.accessioned | 2019-09-17T01:12:22Z | |
dc.date.available | 2019-09-17T01:12:22Z | |
dc.date.issued | 2019 | |
dc.identifier.uri | http://hdl.handle.net/10150/634249 | |
dc.description.abstract | This thesis investigates methods for Acoustic Modeling in Automatic Speech Recog- nition, assuming limited access to training data in the target domain. The Acoustic Models of interest are Deep Neural Network Acoustic Models (in both the Hybrid and End-to-End approaches), and the target domains in question are either different languages or different speakers. Inductive bias is transfered from a source domain during training, via Multi-Task Learning or Transfer Learning. With regards to Multi-Task Learning, Chapter (5) presents experiments which explicitly incorporate linguistic knowledge (i.e. phonetics and phonology) into an auxiliary task during neural Acoustic Model training. In Chapter (6), I investigate Multi-Task methods which do not rely on expert knowledge (linguistic or otherwise), by re-using existing parts of the Hybrid training pipeline. In Chapter (7), new tasks are discovered using unsupervised learning. In Chapter (8), using the “copy-paste” Transfer Learning approach, I demonstrate that with an appropriate early-stopping criteria, cross-lingual transfer is possible to both large and small target datasets. The methods and intuitions which rely on linguistic knowledge are of interest to the Speech Recognition practitioner working in low-resource domains. These same sections may be of interest to the theoretical linguist, as a study of the relative import of phonetic categories in classification. To the Machine Learning practitioner, I hope to offer approaches which can be easily ported over to other classification tasks. To the Machine Learning researcher, I hope to inspire new ideas on addressing the small data problem. | |
dc.language.iso | en | |
dc.publisher | The University of Arizona. | |
dc.rights | Copyright © is held by the author. Digital access to this material is made possible by the University Libraries, University of Arizona. Further transmission, reproduction, presentation (such as public display or performance) of protected items is prohibited except with permission of the author. | |
dc.subject | Automatic Speech Recognition | |
dc.subject | Deep Neural Networks | |
dc.subject | Low-Resource Languages | |
dc.subject | Multi-Task Learning | |
dc.subject | Transfer Learning | |
dc.title | Multi-Task and Transfer Learning in Low-Resource Speech Recognition | |
dc.type | text | |
dc.type | Electronic Dissertation | |
thesis.degree.grantor | University of Arizona | |
thesis.degree.level | doctoral | |
dc.contributor.committeemember | Hammond, Michael | |
dc.contributor.committeemember | Morrison, Clayton | |
dc.description.release | Release after 08/20/2020 | |
thesis.degree.discipline | Graduate College | |
thesis.degree.discipline | Linguistics | |
thesis.degree.name | Ph.D. |