Publication: HiSEG: Human assisted instance segmentation
dc.contributor.department | Department of Computer Engineering | |
dc.contributor.kuauthor | Sezgin, Tevfik Metin | |
dc.contributor.other | Department of Computer Engineering | |
dc.contributor.schoolcollegeinstitute | College of Engineering | |
dc.contributor.schoolcollegeinstitute | Graduate School of Sciences and Engineering | |
dc.date.accessioned | 2024-12-29T09:41:23Z | |
dc.date.issued | 2024 | |
dc.description.abstract | Instance segmentation is a form of image detection which has a range of applications, such as object refinement, medical image analysis, and image/video editing, all of which demand a high degree of accuracy. However, this precision is often beyond the reach of what even state-of-the-art, fully automated instance segmentation algorithms can deliver. The performance gap becomes particularly prohibitive for small and complex objects. Practitioners typically resort to fully manual annotation, which can be a laborious process. In order to overcome this problem, we propose a novel approach to enable more precise predictions and generate higher-quality segmentation masks for high-curvature, complex and small-scale objects. Our human-assisted segmentation method, HiSEG, augments the existing Strong Mask R-CNN network to incorporate human-specified partial boundaries. We also present a dataset of hand-drawn partial object boundaries, which we refer to as “human attention maps”. In addition, the Partial Sketch Object Boundaries (PSOB) dataset contains hand-drawn partial object boundaries which represent curvatures of an object's ground truth mask with several pixels. Through extensive evaluation using the PSOB dataset, we show that HiSEG outperforms state-of-the art methods such as Mask R-CNN, Strong Mask R-CNN, Mask2Former, and Segment Anything, achieving respective increases of +42.0, +34.9, +29.9, and +13.4 points in APMask metrics for these four models. We hope that our novel approach will set a baseline for future human-aided deep learning models by combining fully automated and interactive instance segmentation architectures. | |
dc.description.indexedby | WoS | |
dc.description.indexedby | Scopus | |
dc.description.publisherscope | International | |
dc.description.volume | 124 | |
dc.identifier.doi | 10.1016/j.cag.2024.104061 | |
dc.identifier.eissn | 1873-7684 | |
dc.identifier.issn | 0097-8493 | |
dc.identifier.quartile | Q2 | |
dc.identifier.scopus | 2-s2.0-85202354327 | |
dc.identifier.uri | https://doi.org/10.1016/j.cag.2024.104061 | |
dc.identifier.uri | https://hdl.handle.net/20.500.14288/23620 | |
dc.identifier.wos | 1305644400001 | |
dc.keywords | Data fusion | |
dc.keywords | Human-aided computer vision | |
dc.keywords | Instance segmentation | |
dc.keywords | Interactive segmentation | |
dc.keywords | Pen-based partial annotation | |
dc.language | en | |
dc.publisher | Elsevier Ltd | |
dc.source | Computers and Graphics | |
dc.subject | Computer engineering | |
dc.title | HiSEG: Human assisted instance segmentation | |
dc.type | Journal article | |
dspace.entity.type | Publication | |
local.contributor.kuauthor | Sezgin, Tevfik Metin | |
local.contributor.kuauthor | Korkmaz, Muhammed | |
relation.isOrgUnitOfPublication | 89352e43-bf09-4ef4-82f6-6f9d0174ebae | |
relation.isOrgUnitOfPublication.latestForDiscovery | 89352e43-bf09-4ef4-82f6-6f9d0174ebae |