Publication:
Using synthetic data for person tracking under adverse weather conditions

Placeholder

Organizational Units

Program

KU-Authors

KU Authors

Co-Authors

Kerim, Abdulrahman
Çelikcan, Ufuk
Erdem, Erkut

Advisor

Publication Date

2021

Language

English

Type

Journal Article

Journal Title

Journal ISSN

Volume Title

Abstract

Robust visual tracking plays a vital role in many areas such as autonomous cars, surveillance and robotics. Recent trackers were shown to achieve adequate results under normal tracking scenarios with clear weather conditions, standard camera setups and lighting conditions. Yet, the performance of these trackers, whether they are corre-lation filter-based or learning-based, degrade under adverse weather conditions. The lack of videos with such weather conditions, in the available visual object tracking datasets, is the prime issue behind the low perfor-mance of the learning-based tracking algorithms. In this work, we provide a new person tracking dataset of real-world sequences (PTAW172Real) captured under foggy, rainy and snowy weather conditions to assess the performance of the current trackers. We also introduce a novel person tracking dataset of synthetic sequences (PTAW217Synth) procedurally generated by our NOVA framework spanning the same weather conditions in varying severity to mitigate the problem of data scarcity. Our experimental results demonstrate that the perfor-mances of the state-of-the-art deep trackers under adverse weather conditions can be boosted when the avail-able real training sequences are complemented with our synthetically generated dataset during training. (c) 2021 Elsevier B.V. All rights reserved.

Description

Source:

Image and Vision Computing

Publisher:

Elsevier

Keywords:

Subject

Computer science, Artificial intelligence, Engineering, Software engineering, Theory methods, Engineering, Electrical electronic engineering, Optics

Citation

Endorsement

Review

Supplemented By

Referenced By

Copy Rights Note

0

Views

0

Downloads

View PlumX Details