Publication: Video frame prediction via deep learning
dc.contributor.department | N/A | |
dc.contributor.department | Department of Electrical and Electronics Engineering | |
dc.contributor.kuauthor | Yılmaz, Mustafa Akın | |
dc.contributor.kuauthor | Tekalp, Ahmet Murat | |
dc.contributor.kuprofile | PhD Student | |
dc.contributor.kuprofile | Faculty Member | |
dc.contributor.other | Department of Electrical and Electronics Engineering | |
dc.contributor.schoolcollegeinstitute | Graduate School of Sciences and Engineering | |
dc.contributor.schoolcollegeinstitute | College of Engineering | |
dc.contributor.yokid | N/A | |
dc.contributor.yokid | 26207 | |
dc.date.accessioned | 2024-11-09T23:13:11Z | |
dc.date.issued | 2020 | |
dc.description.abstract | This paper provides new results over our previous work presented in ICIP 2019 on the performance of learned frame prediction architectures and associated training methods. More specifically, we show that using an end-to-end residual connection in the fully convolutional neural network (FCNN) provides improved performance. in order to provide comparative results, we trained a residual FCNN, A convolutional RNN (CRNN), and a convolutional long-short term memory (CLSTM) network for next frame prediction using the mean square loss. We performed both stateless and stateful training for recurrent networks. Experimental results show that the residual FCNN architecture performs the best in terms of peak signal to noise ratio (PSNR) at the expense of higher training and test (inference) computational complexity. the CRNN can be stably and efficiently trained using the stateful truncated backpropagation through time procedure, and requires an order of magnitude less inference runtime to achieve an acceptable performance in near real-time. | |
dc.description.indexedby | WoS | |
dc.description.openaccess | NO | |
dc.description.publisherscope | International | |
dc.description.sponsoredbyTubitakEu | TÜBİTAK | |
dc.description.sponsorship | TUBITAKproject [217E033] | |
dc.description.sponsorship | Turkish academy of Sciences (TUBa) This work was supported by TUBITAKproject 217E033. a. Murat Tekalp also acknowledges support from Turkish academy of Sciences (TUBa). | |
dc.identifier.doi | N/A | |
dc.identifier.isbn | 978-1-7281-7206-4 | |
dc.identifier.issn | 2165-0608 | |
dc.identifier.quartile | N/A | |
dc.identifier.uri | https://hdl.handle.net/20.500.14288/9946 | |
dc.identifier.wos | 653136100021 | |
dc.keywords | Frame prediction | |
dc.keywords | Deep learning | |
dc.keywords | Recurrent network architectures | |
dc.keywords | Stateful training | |
dc.keywords | Convolutional network architectures | |
dc.language | Turkish | |
dc.publisher | IEEE | |
dc.source | 2020 28th Signal Processing and Communications Applications Conference (Siu) | |
dc.subject | Civil engineering | |
dc.subject | Electrical electronics engineering | |
dc.subject | Telecommunication | |
dc.title | Video frame prediction via deep learning | |
dc.type | Conference proceeding | |
dspace.entity.type | Publication | |
local.contributor.authorid | 0000-0002-0795-8970 | |
local.contributor.authorid | 0000-0003-1465-8121 | |
local.contributor.kuauthor | Yılmaz, Mustafa Akın | |
local.contributor.kuauthor | Tekalp, Ahmet Murat | |
relation.isOrgUnitOfPublication | 21598063-a7c5-420d-91ba-0cc9b2db0ea0 | |
relation.isOrgUnitOfPublication.latestForDiscovery | 21598063-a7c5-420d-91ba-0cc9b2db0ea0 |