A Simple Method on Generating Synthetic Data for Training Real-time Object Detection Networks

Jungwoo Huh, Kyoungoh Lee, Inwoong Lee, Sanghoon Lee

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Environment recognition has been an important topic ever since the emergence of augmented reality (AR). For better experience in AR applications, environment recognition should be provided fast in real-time, where real-time object detection technologies could fulfill this requirement. However, training object detectors for AR specific scenarios are often troublesome. The real-time nature of AR produces visual degradations such as motion blur or occlusion by interaction, which make detectors trained with plain data difficult to detect objects exposed in such complex situations. Also, since gathering and labeling training data from scratch is a heavy burden, we need to resort to synthesized training data but previous synthetic data generation frameworks do not consider the aforementioned issue. Therefore, in this paper, we propose a new synthetic data generation framework which includes visual variations such as motion blur and occlusion occurred by distractors. By this simple modification, we show that including such variated data to the training dataset could dramatically improve realtime performance of object detectors by a high margin. Also, we stress that synthesizing training data with no more than three objects per image can achieve competitive performance compared to detectors trained with over four present in a single image. Experimental results both quantitatively and qualitatively supports our statements and shows the superiority of our method.

Original languageEnglish
Title of host publication2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, APSIPA ASC 2018 - Proceedings
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages1518-1522
Number of pages5
ISBN (Electronic)9789881476852
DOIs
Publication statusPublished - 2019 Mar 4
Event10th Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, APSIPA ASC 2018 - Honolulu, United States
Duration: 2018 Nov 122018 Nov 15

Publication series

Name2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, APSIPA ASC 2018 - Proceedings

Conference

Conference10th Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, APSIPA ASC 2018
CountryUnited States
CityHonolulu
Period18/11/1218/11/15

Fingerprint

Augmented reality
Detectors
Labeling
Degradation
Object detection

All Science Journal Classification (ASJC) codes

  • Information Systems

Cite this

Huh, J., Lee, K., Lee, I., & Lee, S. (2019). A Simple Method on Generating Synthetic Data for Training Real-time Object Detection Networks. In 2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, APSIPA ASC 2018 - Proceedings (pp. 1518-1522). [8659778] (2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, APSIPA ASC 2018 - Proceedings). Institute of Electrical and Electronics Engineers Inc.. https://doi.org/10.23919/APSIPA.2018.8659778
Huh, Jungwoo ; Lee, Kyoungoh ; Lee, Inwoong ; Lee, Sanghoon. / A Simple Method on Generating Synthetic Data for Training Real-time Object Detection Networks. 2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, APSIPA ASC 2018 - Proceedings. Institute of Electrical and Electronics Engineers Inc., 2019. pp. 1518-1522 (2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, APSIPA ASC 2018 - Proceedings).
@inproceedings{ed51f5a3f7174d3dadb04761f09bd37a,
title = "A Simple Method on Generating Synthetic Data for Training Real-time Object Detection Networks",
abstract = "Environment recognition has been an important topic ever since the emergence of augmented reality (AR). For better experience in AR applications, environment recognition should be provided fast in real-time, where real-time object detection technologies could fulfill this requirement. However, training object detectors for AR specific scenarios are often troublesome. The real-time nature of AR produces visual degradations such as motion blur or occlusion by interaction, which make detectors trained with plain data difficult to detect objects exposed in such complex situations. Also, since gathering and labeling training data from scratch is a heavy burden, we need to resort to synthesized training data but previous synthetic data generation frameworks do not consider the aforementioned issue. Therefore, in this paper, we propose a new synthetic data generation framework which includes visual variations such as motion blur and occlusion occurred by distractors. By this simple modification, we show that including such variated data to the training dataset could dramatically improve realtime performance of object detectors by a high margin. Also, we stress that synthesizing training data with no more than three objects per image can achieve competitive performance compared to detectors trained with over four present in a single image. Experimental results both quantitatively and qualitatively supports our statements and shows the superiority of our method.",
author = "Jungwoo Huh and Kyoungoh Lee and Inwoong Lee and Sanghoon Lee",
year = "2019",
month = "3",
day = "4",
doi = "10.23919/APSIPA.2018.8659778",
language = "English",
series = "2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, APSIPA ASC 2018 - Proceedings",
publisher = "Institute of Electrical and Electronics Engineers Inc.",
pages = "1518--1522",
booktitle = "2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, APSIPA ASC 2018 - Proceedings",
address = "United States",

}

Huh, J, Lee, K, Lee, I & Lee, S 2019, A Simple Method on Generating Synthetic Data for Training Real-time Object Detection Networks. in 2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, APSIPA ASC 2018 - Proceedings., 8659778, 2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, APSIPA ASC 2018 - Proceedings, Institute of Electrical and Electronics Engineers Inc., pp. 1518-1522, 10th Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, APSIPA ASC 2018, Honolulu, United States, 18/11/12. https://doi.org/10.23919/APSIPA.2018.8659778

A Simple Method on Generating Synthetic Data for Training Real-time Object Detection Networks. / Huh, Jungwoo; Lee, Kyoungoh; Lee, Inwoong; Lee, Sanghoon.

2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, APSIPA ASC 2018 - Proceedings. Institute of Electrical and Electronics Engineers Inc., 2019. p. 1518-1522 8659778 (2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, APSIPA ASC 2018 - Proceedings).

Research output: Chapter in Book/Report/Conference proceedingConference contribution

TY - GEN

T1 - A Simple Method on Generating Synthetic Data for Training Real-time Object Detection Networks

AU - Huh, Jungwoo

AU - Lee, Kyoungoh

AU - Lee, Inwoong

AU - Lee, Sanghoon

PY - 2019/3/4

Y1 - 2019/3/4

N2 - Environment recognition has been an important topic ever since the emergence of augmented reality (AR). For better experience in AR applications, environment recognition should be provided fast in real-time, where real-time object detection technologies could fulfill this requirement. However, training object detectors for AR specific scenarios are often troublesome. The real-time nature of AR produces visual degradations such as motion blur or occlusion by interaction, which make detectors trained with plain data difficult to detect objects exposed in such complex situations. Also, since gathering and labeling training data from scratch is a heavy burden, we need to resort to synthesized training data but previous synthetic data generation frameworks do not consider the aforementioned issue. Therefore, in this paper, we propose a new synthetic data generation framework which includes visual variations such as motion blur and occlusion occurred by distractors. By this simple modification, we show that including such variated data to the training dataset could dramatically improve realtime performance of object detectors by a high margin. Also, we stress that synthesizing training data with no more than three objects per image can achieve competitive performance compared to detectors trained with over four present in a single image. Experimental results both quantitatively and qualitatively supports our statements and shows the superiority of our method.

AB - Environment recognition has been an important topic ever since the emergence of augmented reality (AR). For better experience in AR applications, environment recognition should be provided fast in real-time, where real-time object detection technologies could fulfill this requirement. However, training object detectors for AR specific scenarios are often troublesome. The real-time nature of AR produces visual degradations such as motion blur or occlusion by interaction, which make detectors trained with plain data difficult to detect objects exposed in such complex situations. Also, since gathering and labeling training data from scratch is a heavy burden, we need to resort to synthesized training data but previous synthetic data generation frameworks do not consider the aforementioned issue. Therefore, in this paper, we propose a new synthetic data generation framework which includes visual variations such as motion blur and occlusion occurred by distractors. By this simple modification, we show that including such variated data to the training dataset could dramatically improve realtime performance of object detectors by a high margin. Also, we stress that synthesizing training data with no more than three objects per image can achieve competitive performance compared to detectors trained with over four present in a single image. Experimental results both quantitatively and qualitatively supports our statements and shows the superiority of our method.

UR - http://www.scopus.com/inward/record.url?scp=85063439166&partnerID=8YFLogxK

UR - http://www.scopus.com/inward/citedby.url?scp=85063439166&partnerID=8YFLogxK

U2 - 10.23919/APSIPA.2018.8659778

DO - 10.23919/APSIPA.2018.8659778

M3 - Conference contribution

AN - SCOPUS:85063439166

T3 - 2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, APSIPA ASC 2018 - Proceedings

SP - 1518

EP - 1522

BT - 2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, APSIPA ASC 2018 - Proceedings

PB - Institute of Electrical and Electronics Engineers Inc.

ER -

Huh J, Lee K, Lee I, Lee S. A Simple Method on Generating Synthetic Data for Training Real-time Object Detection Networks. In 2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, APSIPA ASC 2018 - Proceedings. Institute of Electrical and Electronics Engineers Inc. 2019. p. 1518-1522. 8659778. (2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, APSIPA ASC 2018 - Proceedings). https://doi.org/10.23919/APSIPA.2018.8659778