Publication: A deep learning approach for data driven vocal tract area function estimation
dc.contributor.department | Department of Computer Engineering | |
dc.contributor.department | Department of Electrical and Electronics Engineering | |
dc.contributor.kuauthor | Erzin, Engin | |
dc.contributor.kuauthor | Asadiabadi, Sasan | |
dc.contributor.kuprofile | Faculty Member | |
dc.contributor.other | Department of Computer Engineering | |
dc.contributor.other | Department of Electrical and Electronics Engineering | |
dc.contributor.schoolcollegeinstitute | College of Sciences | |
dc.contributor.schoolcollegeinstitute | Graduate School of Sciences and Engineering | |
dc.contributor.yokid | 34503 | |
dc.contributor.yokid | N/A | |
dc.date.accessioned | 2024-11-09T13:45:07Z | |
dc.date.issued | 2018 | |
dc.description.abstract | In this paper we present a data driven vocal tract area function (VTAF) estimation using Deep Neural Networks (DNN). We approach the VTAF estimation problem based on sequence to sequence learning neural networks, where regression over a sliding window is used to learn arbitrary non-linear one-to-many mapping from the input feature sequence to the target articulatory sequence. We propose two schemes for efficient estimation of the VTAF; (1) a direct estimation of the area function values and (2) an indirect estimation via predicting the vocal tract boundaries. We consider acoustic speech and phone sequence as two possible input modalities for the DNN estimators. Experimental evaluations are performed over a large data comprising acoustic and phonetic features with parallel articulatory information from the USC-TIMIT database. Our results show that the proposed direct and indirect schemes perform the VTAF estimation with mean absolute error (MAE) rates lower than 1.65 mm, where the direct estimation scheme is observed to perform better than the indirect scheme. | |
dc.description.fulltext | YES | |
dc.description.indexedby | WoS | |
dc.description.indexedby | Scopus | |
dc.description.openaccess | YES | |
dc.description.publisherscope | International | |
dc.description.sponsoredbyTubitakEu | N/A | |
dc.description.sponsorship | N/A | |
dc.description.version | Author's final manuscript | |
dc.format | ||
dc.identifier.doi | 10.1109/SLT.2018.8639582 | |
dc.identifier.embargo | NO | |
dc.identifier.filenameinventoryno | IR01885 | |
dc.identifier.isbn | 9781538643341 | |
dc.identifier.issn | 2639-5479 | |
dc.identifier.link | https://doi.org/10.1109/SLT.2018.8639582 | |
dc.identifier.quartile | N/A | |
dc.identifier.scopus | 2-s2.0-85063083027 | |
dc.identifier.uri | https://hdl.handle.net/20.500.14288/3581 | |
dc.identifier.wos | 463141800025 | |
dc.keywords | Speech articulation | |
dc.keywords | Vocal tract area function | |
dc.keywords | Deep neural network | |
dc.keywords | Convolutional neural network | |
dc.language | English | |
dc.publisher | Institute of Electrical and Electronics Engineers (IEEE) | |
dc.relation.grantno | NA | |
dc.relation.uri | http://cdm21054.contentdm.oclc.org/cdm/ref/collection/IR/id/8568 | |
dc.source | 2018 IEEE Workshop on Spoken Language Technology (SLT) | |
dc.subject | Computer science | |
dc.subject | Artificial intelligence | |
dc.subject | Engineering, electrical and electronic | |
dc.title | A deep learning approach for data driven vocal tract area function estimation | |
dc.type | Journal Article | |
dspace.entity.type | Publication | |
local.contributor.authorid | 0000-0002-2715-2368 | |
local.contributor.authorid | N/A | |
local.contributor.kuauthor | Erzin, Engin | |
local.contributor.kuauthor | Asadiabadi, Sasan | |
relation.isOrgUnitOfPublication | 89352e43-bf09-4ef4-82f6-6f9d0174ebae | |
relation.isOrgUnitOfPublication | 21598063-a7c5-420d-91ba-0cc9b2db0ea0 | |
relation.isOrgUnitOfPublication.latestForDiscovery | 21598063-a7c5-420d-91ba-0cc9b2db0ea0 |
Files
Original bundle
1 - 1 of 1