Publication:
A gated fusion network for dynamic saliency prediction

dc.contributor.coauthorKocak, Aysun
dc.contributor.coauthorErdem, Erkut
dc.contributor.departmentDepartment of Computer Engineering
dc.contributor.kuauthorErdem, Aykut
dc.contributor.kuprofileFaculty Member
dc.contributor.otherDepartment of Computer Engineering
dc.contributor.schoolcollegeinstituteCollege of Engineering
dc.contributor.yokid20331
dc.date.accessioned2024-11-09T11:36:23Z
dc.date.issued2022
dc.description.abstractPredicting saliency in videos is a challenging problem due to complex modeling of interactions between spatial and temporal information, especially when ever-changing, dynamic nature of videos is considered. Recently, researchers have proposed large-scale data sets and models that take advantage of deep learning as a way to understand what is important for video saliency. These approaches, however, learn to combine spatial and temporal features in a static manner and do not adapt themselves much to the changes in the video content. In this article, we introduce the gated fusion network for dynamic saliency (GFSalNet), the first deep saliency model capable of making predictions in a dynamic way via the gated fusion mechanism. Moreover, our model also exploits spatial and channelwise attention within a multiscale architecture that further allows for highly accurate predictions. We evaluate the proposed approach on a number of data sets, and our experimental analysis demonstrates that it outperforms or is highly competitive with the state of the art. Importantly, we show that it has a good generalization ability, and moreover, exploits temporal information more effectively via its adaptive fusion scheme.
dc.description.fulltextYES
dc.description.indexedbyWoS
dc.description.indexedbyScopus
dc.description.issue3
dc.description.openaccessYES
dc.description.publisherscopeInternational
dc.description.sponsoredbyTubitakEuN/A
dc.description.sponsorshipTurkish Academy of Sciences
dc.description.sponsorshipGEBIP 2018
dc.description.sponsorshipScience Academy
dc.description.sponsorshipBAGEP 2021
dc.description.versionAuthor's final manuscript
dc.description.volume14
dc.formatpdf
dc.identifier.doi10.1109/TCDS.2021.3094974
dc.identifier.eissn2379-8939
dc.identifier.embargoNO
dc.identifier.filenameinventorynoIR04003
dc.identifier.issn2379-8920
dc.identifier.linkhttps://doi.org/10.1109/TCDS.2021.3094974
dc.identifier.quartileQ2
dc.identifier.scopus2-s2.0-85112629359
dc.identifier.urihttps://hdl.handle.net/20.500.14288/53
dc.identifier.wos852243600022
dc.keywordsVideos
dc.keywordsAdaptation models
dc.keywordsPredictive models
dc.keywordsVisualization
dc.keywordsDynamics
dc.keywordsFeature extraction
dc.keywordsLogic gates
dc.keywordsDeep saliency networks
dc.keywordsDynamic saliency estimation
dc.keywordsGated fusion
dc.languageEnglish
dc.publisherInstitute of Electrical and Electronics Engineers (IEEE)
dc.relation.grantnoNA
dc.relation.urihttp://cdm21054.contentdm.oclc.org/cdm/ref/collection/IR/id/10889
dc.sourceIEEE Transactions on Cognitive and Developmental Systems
dc.subjectComputer science
dc.subjectRobotics
dc.subjectNeurosciences and neurology
dc.titleA gated fusion network for dynamic saliency prediction
dc.typeJournal Article
dspace.entity.typePublication
local.contributor.authorid0000-0002-6280-8422
local.contributor.kuauthorErdem, Aykut
relation.isOrgUnitOfPublication89352e43-bf09-4ef4-82f6-6f9d0174ebae
relation.isOrgUnitOfPublication.latestForDiscovery89352e43-bf09-4ef4-82f6-6f9d0174ebae

Files

Original bundle

Now showing 1 - 1 of 1
Thumbnail Image
Name:
10889.pdf
Size:
2.13 MB
Format:
Adobe Portable Document Format