Comparison of Transfer Style Using a CycleGAN Model with Data Augmentation



Título del documento: Comparison of Transfer Style Using a CycleGAN Model with Data Augmentation
Revista: Computación y sistemas
Base de datos:
Número de sistema: 000607843
ISSN: 1405-5546
Autores: 1
1
1
1
Instituciones: 1Instituto Politécnico Nacional, Centro de Investigación en Computación, México
Año:
Periodo: Oct-Dic
Volumen: 27
Número: 4
Paginación: 941-953
País: México
Idioma: Inglés
Resumen en inglés Image-to-image translation (I2I) is a specialized technique aimed at converting images from one domain to another while retaining their intrinsic content. This process involves learning the relationship between an input and its corresponding output image through a dataset of aligned pairs. Our study utilizes the CycleGAN model to pioneer a method for transforming images from the domain of Monet’s paintings to a domain of varied photographs without the need for paired training examples. We address challenges such as mode collapse and overfitting, which can affect the integrity and quality of the translated images. Our investigation focuses on enhancing the CycleGAN model’s performance and stability through data augmentation strategies, such as flipping, mirroring, and contrast enhancement. We propose that judicious dataset selection for training can yield superior outcomes with less data compared to indiscriminate large-volume training. By online scraping Monet’s artwork and curating a diverse, representative image subset, we fine-tuned our model. This targeted approach propelled our results to 2nd place in the Kaggle challenge ”I am something of a Painter Myself” as of August 3rd, 2023, demonstrating the efficacy of our enhanced training protocol.
Keyword: Generative adversarial network,
Image-to-image translation,
Data augmentation,
Cycle consistency
Texto completo: Texto completo (Ver PDF) Texto completo (Ver HTML)