Publication:
Deep learning for image/video restoration and super-resolution

dc.contributor.departmentDepartment of Electrical and Electronics Engineering
dc.contributor.kuauthorTekalp, Ahmet Murat
dc.contributor.schoolcollegeinstituteCollege of Engineering
dc.date.accessioned2024-11-09T13:12:37Z
dc.date.issued2022
dc.description.abstractRecent advances in neural signal processing led to significant improvements in the performance of learned image/video restoration and super-resolution (SR). An important benefit of data-driven deep learning approaches to image processing is that neural models can be optimized for any differentiable loss function, including perceptual loss functions, leading to perceptual image/video restoration and SR, which cannot be easily handled by traditional model-based methods. We start with a brief problem statement and a short discussion on traditional vs. data-driven solutions. We next review recent advances in neural architectures, such as residual blocks, dense connections, residual-in-residual dense blocks, residual blocks with generative neurons, self-attention and visual transformers. We then discuss loss functions and evaluation (assessment) criteria for image/video restoration and SR, including fidelity (distortion) and perceptual criteria, and the relation between them, where we briefly review the perception vs. distortion trade-off. We can consider learned image/video restoration and SR as learning either a nonlinear regressive mapping from degraded to ideal images based on the universal approximation theorem, or a generative model that captures the probability distribution of ideal images. We first review regressive inference via residual and/or dense convolutional networks (ConvNet). We also show that using a new architecture with residual blocks based on a generative neuron model can outperform classical residual ConvNets in peak-signal-to-noise ratio (PSNR). We next discuss generative inference based on adversarial training, such as SRGAN and ESRGAN, which can reproduce realistic textures, or based on normalizing flow such as SRFlow by optimizing log-likelihood. We then discuss problems in applying supervised training to real-life restoration and SR, including overfitting image priors and overfitting the degradation model seen in the training set. We introduce multiple-model SR and real-world SR (from unpaired training data) formulations to overcome these problems. Integration of traditional model-based methods and deep learning for non-blind restoration/SR is introduced as another solution to model overfitting in supervised learning. In learned video restoration and SR (VSR), we first discuss how to best exploit temporal correlations in video, including sliding temporal window vs. recurrent architectures for propagation, and aligning frames in the pixel domain using optical flow vs. in the feature space using deformable convolutions. We next introduce early fusion with feature-space alignment, employed by the EDVR network, which obtains excellent PSNR performance. However, it is well-known that videos with the highest PSNR may not be the most appealing to humans, since minimizing the mean-square error may result in blurring of details. We then address perceptual optimization of VSR models to obtain natural texture and motion. Although perception-distortion tradeoff has been well studied for images, few works address perceptual VSR. In addition to using perceptual losses, such as MS-SSIM, LPIPS, and/or adversarial training, we also discuss explicit loss functions/criteria to enforce and evaluate temporal consistency. We conclude with a discussion of open problems.
dc.description.fulltextYES
dc.description.indexedbyScopus
dc.description.issue1
dc.description.openaccessYES
dc.description.publisherscopeInternational
dc.description.sponsoredbyTubitakEuTÜBİTAK
dc.description.sponsorshipScientific and Technological Research Council of Turkey (TÜBİTAK)
dc.description.sponsorshipTUBITAK 2247-A
dc.description.versionAuthor's final manuscript
dc.description.volume13
dc.identifier.doi10.1561/0600000100
dc.identifier.embargoNO
dc.identifier.filenameinventorynoIR03706
dc.identifier.issn1572-2740
dc.identifier.quartileN/A
dc.identifier.scopus2-s2.0-85130244443
dc.identifier.urihttps://doi.org/10.1561/0600000100
dc.keywordsBackpropagation
dc.keywordsConvolution
dc.keywordsConvolutional neural networks
dc.keywordsEconomic and social effects
dc.keywordsImage reconstruction
dc.language.isoeng
dc.publisherNow Publisher
dc.relation.grantno120C156
dc.relation.ispartofFoundations and Trends in Computer Graphics and Vision
dc.relation.urihttp://cdm21054.contentdm.oclc.org/cdm/ref/collection/IR/id/10566
dc.subjectEngineering
dc.titleDeep learning for image/video restoration and super-resolution
dc.typeJournal Article
dspace.entity.typePublication
local.contributor.kuauthorTekalp, Ahmet Murat
local.publication.orgunit1College of Engineering
local.publication.orgunit2Department of Electrical and Electronics Engineering
relation.isOrgUnitOfPublication21598063-a7c5-420d-91ba-0cc9b2db0ea0
relation.isOrgUnitOfPublication.latestForDiscovery21598063-a7c5-420d-91ba-0cc9b2db0ea0
relation.isParentOrgUnitOfPublication8e756b23-2d4a-4ce8-b1b3-62c794a8c164
relation.isParentOrgUnitOfPublication.latestForDiscovery8e756b23-2d4a-4ce8-b1b3-62c794a8c164

Files

Original bundle

Now showing 1 - 1 of 1
Thumbnail Image
Name:
10566.pdf
Size:
1.79 MB
Format:
Adobe Portable Document Format