GAN-Based Image Colorization for Self-Supervised Visual Feature Learning.

Sandra Treneska, Eftim Zdravevski, Ivan Miguel Pires, Petre Lameski, Sonja Gievska
Author Information
  1. Sandra Treneska: Faculty of Computer Science and Engineering, University Ss. Cyril and Methodius, 1000 Skopje, North Macedonia. ORCID
  2. Eftim Zdravevski: Faculty of Computer Science and Engineering, University Ss. Cyril and Methodius, 1000 Skopje, North Macedonia. ORCID
  3. Ivan Miguel Pires: Instituto de Telecomunicações, Universidade da Beira Interior, 6200-001 Covilhã, Portugal. ORCID
  4. Petre Lameski: Faculty of Computer Science and Engineering, University Ss. Cyril and Methodius, 1000 Skopje, North Macedonia. ORCID
  5. Sonja Gievska: Faculty of Computer Science and Engineering, University Ss. Cyril and Methodius, 1000 Skopje, North Macedonia. ORCID

Abstract

Large-scale labeled datasets are generally necessary for successfully training a deep neural network in the computer vision domain. In order to avoid the costly and tedious work of manually annotating image datasets, self-supervised learning methods have been proposed to learn general visual features automatically. In this paper, we first focus on image colorization with generative adversarial networks (GANs) because of their ability to generate the most realistic colorization results. Then, via transfer learning, we use this as a proxy task for visual understanding. Particularly, we propose to use conditional GANs (cGANs) for image colorization and transfer the gained knowledge to two other downstream tasks, namely, multilabel image classification and semantic segmentation. This is the first time that GANs have been used for self-supervised feature learning through image colorization. Through extensive experiments with the COCO and Pascal datasets, we show an increase of 5% for the classification task and 2.5% for the segmentation task. This demonstrates that image colorization with conditional GANs can boost other downstream tasks' performance without the need for manual annotation.

Keywords

References

  1. IEEE Trans Pattern Anal Mach Intell. 2017 Jun;39(6):1137-1149 [PMID: 27295650]
  2. Sci Rep. 2019 Aug 12;9(1):11591 [PMID: 31406194]
  3. IEEE Trans Pattern Anal Mach Intell. 2021 Nov;43(11):4037-4058 [PMID: 32386141]
  4. IEEE Trans Pattern Anal Mach Intell. 2018 Jun;40(6):1452-1464 [PMID: 28692961]
  5. IEEE Trans Image Process. 2004 Apr;13(4):600-12 [PMID: 15376593]
  6. IEEE Trans Pattern Anal Mach Intell. 2018 Apr;40(4):834-848 [PMID: 28463186]

Grants

  1. UIDB/50008/2020/FEDER---PT2020 partnership agreement
  2. FCT/MEC/FCT/MEC through national funds

MeSH Term

Computers
Image Processing, Computer-Assisted
Neural Networks, Computer
Semantics
Vision, Ocular

Word Cloud

Created with Highcharts 10.0.0imagecolorizationlearningGANsdatasetsnetworkself-supervisedtransfertaskneuralvisualfirstgenerativeadversarialuseconditionaldownstreamclassificationsegmentation5%Large-scalelabeledgenerallynecessarysuccessfullytrainingdeepcomputervisiondomainorderavoidcostlytediousworkmanuallyannotatingmethodsproposedlearngeneralfeaturesautomaticallypaperfocusnetworksabilitygeneraterealisticresultsviaproxyunderstandingParticularlyproposecGANsgainedknowledgetwotasksnamelymultilabelsemantictimeusedfeatureextensiveexperimentsCOCOPascalshowincrease2demonstratescanboosttasks'performancewithoutneedmanualannotationGAN-BasedImageColorizationSelf-SupervisedVisualFeatureLearningconvolutional

Similar Articles

Cited By