Publication:
NOVA: rendering virtual worlds with humans for computer vision tasks

dc.contributor.coauthorKerim, Abdulrahman
dc.contributor.coauthorAslan, Cem
dc.contributor.coauthorÇelikcan, Ufuk
dc.contributor.coauthorErdem, Erkut
dc.contributor.departmentDepartment of Computer Engineering
dc.contributor.kuauthorErdem, Aykut
dc.contributor.schoolcollegeinstituteCollege of Engineering
dc.date.accessioned2024-11-09T23:36:10Z
dc.date.issued2021
dc.description.abstractToday, the cutting edge of computer vision research greatly depends on the availability of large datasets, which are critical for effectively training and testing new methods. Manually annotating visual data, however, is not only a labor-intensive process but also prone to errors. In this study, we present NOVA, a versatile framework to create realistic-looking 3D rendered worlds containing procedurally generated humans with rich pixel-level ground truth annotations. NOVA can simulate various environmental factors such as weather conditions or different times of day, and bring an exceptionally diverse set of humans to life, each having a distinct body shape, gender and age. To demonstrate NOVA's capabilities, we generate two synthetic datasets for person tracking. The first one includes 108 sequences, each with different levels of difficulty like tracking in crowded scenes or at nighttime and aims for testing the limits of current state-of-the-art trackers. A second dataset of 97 sequences with normal weather conditions is used to show how our synthetic sequences can be utilized to train and boost the performance of deep-learning based trackers. Our results indicate that the synthetic data generated by NOVA represents a good proxy of the real-world and can be exploited for computer vision tasks.
dc.description.indexedbyWOS
dc.description.indexedbyScopus
dc.description.issue6
dc.description.openaccessNO
dc.description.publisherscopeInternational
dc.description.sponsoredbyTubitakEuN/A
dc.description.volume40
dc.identifier.doi10.1111/cgf.14271
dc.identifier.eissn1467-8659
dc.identifier.issn0167-7055
dc.identifier.quartileQ2
dc.identifier.scopus2-s2.0-85105304770
dc.identifier.urihttps://doi.org/10.1111/cgf.14271
dc.identifier.urihttps://hdl.handle.net/20.500.14288/12607
dc.identifier.wos648344400001
dc.keywordsProcedural content generation
dc.keywordsSynthetic‐data for learning
dc.keywordsVisual tracking
dc.language.isoeng
dc.publisherWiley
dc.relation.ispartofComputer Graphics Forum
dc.subjectComputer science
dc.subjectEngineering
dc.subjectSoftware engineering
dc.titleNOVA: rendering virtual worlds with humans for computer vision tasks
dc.typeJournal Article
dspace.entity.typePublication
local.contributor.kuauthorErdem, Aykut
local.publication.orgunit1College of Engineering
local.publication.orgunit2Department of Computer Engineering
relation.isOrgUnitOfPublication89352e43-bf09-4ef4-82f6-6f9d0174ebae
relation.isOrgUnitOfPublication.latestForDiscovery89352e43-bf09-4ef4-82f6-6f9d0174ebae
relation.isParentOrgUnitOfPublication8e756b23-2d4a-4ce8-b1b3-62c794a8c164
relation.isParentOrgUnitOfPublication.latestForDiscovery8e756b23-2d4a-4ce8-b1b3-62c794a8c164

Files