Publication:
Leveraging semantic saliency maps for query-specific video summarization

dc.contributor.coauthorCizmeciler, Kemal
dc.contributor.coauthorErdem, Erkut
dc.contributor.departmentDepartment of Computer Engineering
dc.contributor.kuauthorErdem, Aykut
dc.contributor.kuprofileFaculty Member
dc.contributor.otherDepartment of Computer Engineering
dc.contributor.schoolcollegeinstituteCollege of Engineering
dc.contributor.yokid20331
dc.date.accessioned2024-11-09T22:53:29Z
dc.date.issued2022
dc.description.abstractThe immense amount of videos being uploaded to video sharing platforms makes it impossible for a person to watch all the videos understand what happens in them. Hence, machine learning techniques are now deployed to index videos by recognizing key objects, actions and scenes or places. Summarization is another alternative as it offers to extract only important parts while covering the gist of the video content. Ideally, the user may prefer to analyze a certain action or scene by searching a query term within the video. Current summarization methods generally do not take queries into account or require exhaustive data labeling. In this work, we present a weakly supervised query-focused video summarization method. Our proposed approach makes use of semantic attributes as an indicator of query relevance and semantic attention maps to locate related regions in the frames and utilizes both within a submodular maximization framework. We conducted experiments on the recently introduced RAD dataset and obtained highly competitive results. Moreover, to better evaluate the performance of our approach on longer videos, we collected a new dataset, which consists of 10 videos from YouTube and annotated with shot-level multiple attributes. Our dataset enables much diverse set of queries that can be used to summarize a video from different perspectives with more degrees of freedom.
dc.description.indexedbyWoS
dc.description.indexedbyScopus
dc.description.issue12
dc.description.openaccessNO
dc.description.publisherscopeInternational
dc.description.sponsorshipGEBIP 2018 Award of the Turkish Academy of Sciences
dc.description.sponsorshipBAGEP 2021 Award of the Science Academy This work was supported in part by GEBIP 2018 Award of the Turkish Academy of Sciences to E. Erdem, BAGEP 2021 Award of the Science Academy to A. Erdem.
dc.description.volume81
dc.identifier.doi10.1007/s11042-022-12442-w
dc.identifier.eissn1573-7721
dc.identifier.issn1380-7501
dc.identifier.quartileQ2
dc.identifier.scopus2-s2.0-85125731088
dc.identifier.urihttp://dx.doi.org/10.1007/s11042-022-12442-w
dc.identifier.urihttps://hdl.handle.net/20.500.14288/7197
dc.identifier.wos765701900019
dc.keywordsQuery-specific
dc.keywordsVideo summarization
dc.keywordsEgocentric video
dc.keywordsScience
dc.languageEnglish
dc.publisherSpringer
dc.sourceMultimedia Tools and Applications
dc.subjectComputer science
dc.subjectInformation systems
dc.subjectEngineering
dc.subjectSoftware engineering
dc.subjectTheory methods
dc.subjectEngineering
dc.subjectElectrical electronic engineering
dc.titleLeveraging semantic saliency maps for query-specific video summarization
dc.typeJournal Article
dspace.entity.typePublication
local.contributor.authorid0000-0002-6280-8422
local.contributor.kuauthorErdem, Aykut
relation.isOrgUnitOfPublication89352e43-bf09-4ef4-82f6-6f9d0174ebae
relation.isOrgUnitOfPublication.latestForDiscovery89352e43-bf09-4ef4-82f6-6f9d0174ebae

Files