Senescence Reversion in Plant Images Using Perception and Unpaired Data

Authors

Keywords:

Image-to-image translation, Unpaired data, CycleGAN, Herbaria, Senescence reversion

Abstract

Recent work on using herbarium images for automatic plant identification, and in particular to do domain adaptation to field images, has been promising. A potential way to address such domain adaptation problem is generative: Hallucinate how a herbarium image would have looked like when it was in the field, and use such synthetic, fresh and green image for plant identification purposes. Such generative task, called senescence reversion, has been poorly explored. To our knowledge, it has been studied only in terms of paired data, meaning, with pairs of dry and fresh images of the same specimen. Such paired data is hard to produce, curate and find. In this work we explore herbarium senescence reversion via unpaired data. The lack of pairs at specimen level presents its own challenges, as capturing the intricacies of each species depends on images of different specimens from different domains. We explore learning a mapping from a herbarium image to leaf plant images and vice-versa, aligning two models, one for each herbarium and leaf domain. We experiment against the state-of-the-art paired baseline Pix2Pix which yields a SSIM of 0.8986, compared to our unpaired approach that yields 0.8865, showing very similar results on reconstruction metrics regardless of our approach not having the luxury of pairs by specimen. Additionally, we apply perceptual loss in order to improve the natural look of the synthetic images. The balance of how much perception is good to avoid reconstruction problems is also studied. Lastly, using a new unpaired dataset built by ourselves, our results show that using a low lambda value from 0.025 to 0.05 for perceptual loss, helps getting lower Fráchet Inception Distances and higher Inception Scores. To our knowledge, no other work has focused on reverting senescence of herbarium sheet images based on unpaired data.

Downloads

Download data is not yet available.

Author Biographies

Esteban A. Esquivel-Barboza, Instituto Tecnológico de Costa Rica

Esteban A. Esquivel-Barboza is a MSc in Computer Science student at the Costa Rican Institute of Technology where he received his B.Eng in Computer Engineering. He interned at Universidade da Coruña, Spain, working in computer vision modules for educational smartphone-based robots. He currently works at Ex2 Outcoding as a Machine Learning Engineer.

Jose Carranza-Rojas, Microsoft

Dr. Jose Carranza-Rojas obtained his Doctorate of Engineering at the Costa Rican Institute of Technology with advisorship from INRIA and CIRAD, France on automatic plant identification. He worked as a Research Scientist at Rakuten Institute of Technology, Japan, on image generative modeling. He currently works as an Associate Professor at Costa Rica Institute of Technology and as a Data and Applied Scientist at Microsoft, in topics related to Natural Language Processing, Computer Vision and Biodiversity Informatics.

References

M. Yerman, “Herbarium Explorations,” American Gardener, 2017.

J. Carranza-Rojas, H. Goeau, P. Bonnet, E. MataMontero, and A. Joly, “Going deeper in the automated

identification of Herbarium specimens,” BMC Evolutionary Biology, vol. 17, no. 1, p. 181, Aug. 11, 2017,

ISSN: 1471-2148.

H. Goeau, P. Bonnet, and A. Joly, “Overview of LifeCLEF Plant Identification task 2020,” p. 15, 2020.

M.-Y. Liu, T. Breuel, and J. Kautz, “Unsupervised

Image-to-Image Translation Networks,” Jul. 22, 2018.

arXiv: 1703.00848 [cs].

Y. S. Chen, Y. C. Wang, M. H. Kao, and Y. Y. Chuang,

“Deep Photo Enhancer: Unpaired Learning for Image

Enhancement from Photographs with GANs,” Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 6306–6314,

, ISSN: 9781538664209.

A. Affouard, H. Goeau, P. Bonnet, J.-C. Lombardo, and ¨

A. Joly, “Pl@ntNet app in the era of deep learning,”

p. 7, 2017.

B. R. Hussein, O. A. Malik, W.-H. Ong, and J. W. F.

Slik, “Reconstruction of damaged herbarium leaves

using deep learning techniques for improving classification accuracy,” Ecological Informatics, vol. 61,

pp. 101 243–101 243, Feb. 2021.

J. Villacis-Llobet, M. Lucio-Troya, M. Calvo-Navarro,

S. Calderon-Ramirez, and E. Mata-Montero, “A First

Glance into Reversing Senescence on Herbarium Sample Images Through Conditional Generative Adversarial Networks,” in Communications in Computer and

Information Science, vol. 1087 CCIS, Springer, 2020,

pp. 438–447, ISBN: 978-3-030-41004-9.

J. Y. Zhu, T. Park, P. Isola, and A. A. Efros,

“Unpaired Image-to-Image Translation Using CycleConsistent Adversarial Networks,” 9781538610329,

, pp. 2242–2251.

H. Talebi and P. Milanfar, “NIMA: Neural Image Assessment,” IEEE Transactions on Image Processing,

vol. 27, no. 8, pp. 3998–4011, 2018.

Y. Blau and T. Michaeli, “The Perception-Distortion

Tradeoff,” 2018.

P. Isola, J. Y. Zhu, T. Zhou, and A. A. Efros, “Imageto-image translation with conditional adversarial networks,” 9781538604571, 2017, pp. 5967–5976.

I. J. Goodfellow, J. Pouget-Abadie, M. Mirza, et al.,

“Generative Adversarial Networks,” Jun. 2014.

Z. Wang, A. Bovik, H. Sheikh, and E. Simoncelli,

“Image Quality Assessment: From Error Visibility to

Structural Similarity,” IEEE Transactions on Image Processing, vol. 13, no. 4, pp. 600–612, Apr. 2004, ISSN:

-7149.

O. Ronneberger, P. Fischer, and T. Brox, “U-Net: Convolutional Networks for Biomedical Image Segmentation,” May 18, 2015. arXiv: 1505.04597 [cs].

K. Simonyan and A. Zisserman, “Very Deep Convolutional Networks for Large-Scale Image Recognition,”

Apr. 10, 2015. arXiv: 1409.1556 [cs].

M. Ounsworth. “Anticipatory movement planning for

quadrotor visual servoeing.” in collab. with G. L. Dudek

(Internal/Supervisor). (), [Online]. Available: https : / /

escholarship . mcgill . ca / concern / theses / zc77ss96z

(visited on 05/26/2021).

X. Mao, Q. Li, H. Xie, R. Y. K. Lau, Z. Wang, and

S. P. Smolley, “Least Squares Generative Adversarial

Networks,” Apr. 5, 2017. arXiv: 1611.04076 [cs].

X. Huang, M.-Y. Liu, S. Belongie, and J. Kautz,

“Multimodal Unsupervised Image-to-Image Translation,” Aug. 14, 2018. arXiv: 1804 . 04732 [cs,

stat].

M. Amodio and S. Krishnaswamy, “TraVeLGAN:

Image-to-image Translation by Transformation Vector

Learning,” Feb. 25, 2019. arXiv: 1902.09631 [cs].

T. Wang and Y. Lin, “CycleGAN with Better Cycles,”

p. 8,

T. Park, A. A. Efros, R. Zhang, and J.-Y. Zhu, “Contrastive Learning for Unpaired Image-to-Image Translation,” Aug. 20, 2020. arXiv: 2007.15651 [cs].

T. Chen, S. Kornblith, M. Norouzi, and G. Hinton, “A

Simple Framework for Contrastive Learning of Visual

Representations,” Jun. 30, 2020. arXiv: 2002 . 05709

[cs, stat].

J. Johnson, A. Alahi, and L. Fei-Fei, “Perceptual

losses for real-time style transfer and super-resolution,”

, 2016, pp. 694–711.

R. Zhang, P. Isola, A. A. Efros, E. Shechtman, and O.

Wang, “The Unreasonable Effectiveness of Deep Features as a Perceptual Metric,” 9781538664209, 2018,

pp. 586–595.

J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and

L. Fei-Fei, “ImageNet: A large-scale hierarchical image database,” in 2009 IEEE Conference on Computer

Vision and Pattern Recognition, Jun. 2009, pp. 248–255.

Published

2022-08-30

How to Cite

Esquivel-Barboza, E. A., & Carranza-Rojas, J. (2022). Senescence Reversion in Plant Images Using Perception and Unpaired Data. IEEE Latin America Transactions, 20(11), 2346–2353. Retrieved from https://latamt.ieeer9.org/index.php/transactions/article/view/6765