Publication: A gated fusion network for dynamic saliency prediction
dc.contributor.coauthor | Kocak, Aysun | |
dc.contributor.coauthor | Erdem, Erkut | |
dc.contributor.department | Department of Computer Engineering | |
dc.contributor.kuauthor | Erdem, Aykut | |
dc.contributor.kuprofile | Faculty Member | |
dc.contributor.other | Department of Computer Engineering | |
dc.contributor.schoolcollegeinstitute | College of Engineering | |
dc.contributor.yokid | 20331 | |
dc.date.accessioned | 2024-11-09T11:36:23Z | |
dc.date.issued | 2022 | |
dc.description.abstract | Predicting saliency in videos is a challenging problem due to complex modeling of interactions between spatial and temporal information, especially when ever-changing, dynamic nature of videos is considered. Recently, researchers have proposed large-scale data sets and models that take advantage of deep learning as a way to understand what is important for video saliency. These approaches, however, learn to combine spatial and temporal features in a static manner and do not adapt themselves much to the changes in the video content. In this article, we introduce the gated fusion network for dynamic saliency (GFSalNet), the first deep saliency model capable of making predictions in a dynamic way via the gated fusion mechanism. Moreover, our model also exploits spatial and channelwise attention within a multiscale architecture that further allows for highly accurate predictions. We evaluate the proposed approach on a number of data sets, and our experimental analysis demonstrates that it outperforms or is highly competitive with the state of the art. Importantly, we show that it has a good generalization ability, and moreover, exploits temporal information more effectively via its adaptive fusion scheme. | |
dc.description.fulltext | YES | |
dc.description.indexedby | WoS | |
dc.description.indexedby | Scopus | |
dc.description.issue | 3 | |
dc.description.openaccess | YES | |
dc.description.publisherscope | International | |
dc.description.sponsoredbyTubitakEu | N/A | |
dc.description.sponsorship | Turkish Academy of Sciences | |
dc.description.sponsorship | GEBIP 2018 | |
dc.description.sponsorship | Science Academy | |
dc.description.sponsorship | BAGEP 2021 | |
dc.description.version | Author's final manuscript | |
dc.description.volume | 14 | |
dc.format | ||
dc.identifier.doi | 10.1109/TCDS.2021.3094974 | |
dc.identifier.eissn | 2379-8939 | |
dc.identifier.embargo | NO | |
dc.identifier.filenameinventoryno | IR04003 | |
dc.identifier.issn | 2379-8920 | |
dc.identifier.link | https://doi.org/10.1109/TCDS.2021.3094974 | |
dc.identifier.quartile | Q2 | |
dc.identifier.scopus | 2-s2.0-85112629359 | |
dc.identifier.uri | https://hdl.handle.net/20.500.14288/53 | |
dc.identifier.wos | 852243600022 | |
dc.keywords | Videos | |
dc.keywords | Adaptation models | |
dc.keywords | Predictive models | |
dc.keywords | Visualization | |
dc.keywords | Dynamics | |
dc.keywords | Feature extraction | |
dc.keywords | Logic gates | |
dc.keywords | Deep saliency networks | |
dc.keywords | Dynamic saliency estimation | |
dc.keywords | Gated fusion | |
dc.language | English | |
dc.publisher | Institute of Electrical and Electronics Engineers (IEEE) | |
dc.relation.grantno | NA | |
dc.relation.uri | http://cdm21054.contentdm.oclc.org/cdm/ref/collection/IR/id/10889 | |
dc.source | IEEE Transactions on Cognitive and Developmental Systems | |
dc.subject | Computer science | |
dc.subject | Robotics | |
dc.subject | Neurosciences and neurology | |
dc.title | A gated fusion network for dynamic saliency prediction | |
dc.type | Journal Article | |
dspace.entity.type | Publication | |
local.contributor.authorid | 0000-0002-6280-8422 | |
local.contributor.kuauthor | Erdem, Aykut | |
relation.isOrgUnitOfPublication | 89352e43-bf09-4ef4-82f6-6f9d0174ebae | |
relation.isOrgUnitOfPublication.latestForDiscovery | 89352e43-bf09-4ef4-82f6-6f9d0174ebae |
Files
Original bundle
1 - 1 of 1