Publication: Leveraging semantic saliency maps for query-specific video summarization
dc.contributor.coauthor | Cizmeciler, Kemal | |
dc.contributor.coauthor | Erdem, Erkut | |
dc.contributor.department | Department of Computer Engineering | |
dc.contributor.kuauthor | Erdem, Aykut | |
dc.contributor.kuprofile | Faculty Member | |
dc.contributor.other | Department of Computer Engineering | |
dc.contributor.schoolcollegeinstitute | College of Engineering | |
dc.contributor.yokid | 20331 | |
dc.date.accessioned | 2024-11-09T22:53:29Z | |
dc.date.issued | 2022 | |
dc.description.abstract | The immense amount of videos being uploaded to video sharing platforms makes it impossible for a person to watch all the videos understand what happens in them. Hence, machine learning techniques are now deployed to index videos by recognizing key objects, actions and scenes or places. Summarization is another alternative as it offers to extract only important parts while covering the gist of the video content. Ideally, the user may prefer to analyze a certain action or scene by searching a query term within the video. Current summarization methods generally do not take queries into account or require exhaustive data labeling. In this work, we present a weakly supervised query-focused video summarization method. Our proposed approach makes use of semantic attributes as an indicator of query relevance and semantic attention maps to locate related regions in the frames and utilizes both within a submodular maximization framework. We conducted experiments on the recently introduced RAD dataset and obtained highly competitive results. Moreover, to better evaluate the performance of our approach on longer videos, we collected a new dataset, which consists of 10 videos from YouTube and annotated with shot-level multiple attributes. Our dataset enables much diverse set of queries that can be used to summarize a video from different perspectives with more degrees of freedom. | |
dc.description.indexedby | WoS | |
dc.description.indexedby | Scopus | |
dc.description.issue | 12 | |
dc.description.openaccess | NO | |
dc.description.publisherscope | International | |
dc.description.sponsorship | GEBIP 2018 Award of the Turkish Academy of Sciences | |
dc.description.sponsorship | BAGEP 2021 Award of the Science Academy This work was supported in part by GEBIP 2018 Award of the Turkish Academy of Sciences to E. Erdem, BAGEP 2021 Award of the Science Academy to A. Erdem. | |
dc.description.volume | 81 | |
dc.identifier.doi | 10.1007/s11042-022-12442-w | |
dc.identifier.eissn | 1573-7721 | |
dc.identifier.issn | 1380-7501 | |
dc.identifier.quartile | Q2 | |
dc.identifier.scopus | 2-s2.0-85125731088 | |
dc.identifier.uri | http://dx.doi.org/10.1007/s11042-022-12442-w | |
dc.identifier.uri | https://hdl.handle.net/20.500.14288/7197 | |
dc.identifier.wos | 765701900019 | |
dc.keywords | Query-specific | |
dc.keywords | Video summarization | |
dc.keywords | Egocentric video | |
dc.keywords | Science | |
dc.language | English | |
dc.publisher | Springer | |
dc.source | Multimedia Tools and Applications | |
dc.subject | Computer science | |
dc.subject | Information systems | |
dc.subject | Engineering | |
dc.subject | Software engineering | |
dc.subject | Theory methods | |
dc.subject | Engineering | |
dc.subject | Electrical electronic engineering | |
dc.title | Leveraging semantic saliency maps for query-specific video summarization | |
dc.type | Journal Article | |
dspace.entity.type | Publication | |
local.contributor.authorid | 0000-0002-6280-8422 | |
local.contributor.kuauthor | Erdem, Aykut | |
relation.isOrgUnitOfPublication | 89352e43-bf09-4ef4-82f6-6f9d0174ebae | |
relation.isOrgUnitOfPublication.latestForDiscovery | 89352e43-bf09-4ef4-82f6-6f9d0174ebae |