[1] R. N. Strange and P. R. Scott, “Plant disease: a threat to global food security,” Annual Review
of Phytopathology, vol. 43, pp. 83–116, August 2005.
[2] S. Savary, L. Willocquet, S. J. Pethybridge, P. Esker, N. McRoberts, and A. Nelson, “The global
burden of pathogens and pests on major food crops,” Nature Ecology & Evolution, vol. 3, no. 3,
pp. 430–439, March 2019.
[3] F. Qin, D. Liu, B. Sun, L. Ruan, Z. Ma, and H. Wang, “Identification of alfalfa leaf diseases
using image recognition technology,” PLoS One, vol. 11, no. 12, p. e0168274, December 2016.
[4] C. Cortes and V. Vapnik, “Support-vector networks,” Machine learning, vol. 20, no. 3, pp.
273–297, March 1995.
[5] L. Hallau, M. Neumann, B. Klatt, B. Kleinhenz, T. Klein, C. Kuhn, M. R¨ohrig, C. Bauckhage, K. Kersting, A.-K. Mahlein et al., “Automated identification of sugar beet diseases using
smartphones,” Plant Pathology, vol. 67, no. 2, pp. 399–410, February 2018.
[6] E. Mwebaze and G. Owomugisha, “Machine learning for plant disease incidence and severity
measurements from leaf images,” in Proceedings of the 15th IEEE International Conference on
Machine Learning and Applications, December 2016, pp. 158–163.
[7] E. Rublee, V. Rabaud, K. Konolige, and G. Bradski, “Orb: An efficient alternative to sift or
surf,” in Proceedings of the International Conference on Computer Vision, November 2011, pp.
2564–2571.
[8] Y. Es-saady, I. El Massi, M. El Yassa, D. Mammass, and A. Benazoun, “Automatic recognition
of plant leaves diseases based on serial combination of two svm classifiers,” in Proceedings of the
International Conference on Electrical and Information Technologies, May 2016, pp. 561–566.
[9] L. C. Ngugi, M. Abelwahab, and M. Abo-Zahhad, “Recent advances in image processing techniques for automated leaf pest and disease recognition-a review,” Information Processing in
Agriculture, vol. 8, no. 1, pp. 27–51, March 2021.
[10] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “Imagenet classification with deep convolutional
neural networks,” in Proceedings of the Advances in Neural Information Processing Systems,
vol. 25, December 2012, pp. 1097–1105.
71
[11] J. Deng, W. Dong, R. Socher, L. J. Li, K. Li, and L. Fei-Fei, “Imagenet: A large-scale hierarchical
image database,” in Proceedings of the IEEE Conference on Computer Vision and Pattern
Recognition, June 2009, pp. 248–255.
[12] B. Liu, Y. Zhang, D. He, and Y. Li, “Identification of apple leaf diseases based on deep convolutional neural networks,” Symmetry, vol. 10, no. 1, pp. 1–16, January 2018.
[13] C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke, and
A. Rabinovich, “Going deeper with convolutions,” in Proceedings of the IEEE Conference on
Computer Vision and Pattern Recognition, June 2015, pp. 1–9.
[14] S. P. Mohanty, D. P. Hughes, and M. Salath´e, “Using deep learning for image-based plant disease
detection,” Frontiers in Plant Science, vol. 7, p. 1419, September 2016.
[15] D. Hughes and M. Salath´e, “An open access repository of images on plant health to enable
the development of mobile disease diagnostics,” arXiv:1511.08060, November 2015. [Online].
Available: https://arxiv.org/abs/1511.08060
[16] G. Wang, Y. Sun, and J. Wang, “Automatic image-based plant disease severity estimation using
deep learning,” Computational Intelligence and Neuroscience, vol. 2017, p. 2917536, July 2017.
[17] K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” in Proceedings of the International Conference on Learning Representations, May 2015,
pp. 1–14.
[18] H. Durmu¸s, E. O. G¨
une¸s, and M. Kırcı, “Disease detection on the leaves of the tomato
plants by using deep learning,” in Proceedings of the 6th International Conference on AgroGeoinformatics, August 2017, pp. 1–5.
[19] F. N. Iandola, S. Han, M. W. Moskewicz, K. Ashraf, W. J. Dally, and K. Keutzer, “Squeezenet:
Alexnet-level accuracy with 50x fewer parameters and¡ 0.5 mb model size,” in Proceedings of
the International Conference on Learning Representations, April 2017, pp. 1–13.
[20] A. Elhassouny and F. Smarandache, “Smart mobile application to recognize tomato leaf diseases using convolutional neural networks,” in Proceedings of the International Conference of
Computer Science and Renewable Energies, July 2019, pp. 1–4.
72
[21] A. G. Howard, M. Zhu, B. Chen, D. Kalenichenko, W. Wang, T. Weyand, M. Andreetto, and
H. Adam, “Mobilenets: Efficient convolutional neural networks for mobile vision applications,”
arXiv:1704.04861, April 2017. [Online]. Available: https://arxiv.org/abs/1704.04861
[22] H. A. Atabay, “Deep residual learning for tomato plant leaf disease identification,” Journal of
Theoretical & Applied Information Technology, vol. 95, pp. 6800–6808, December 2017.
[23] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, June 2016, pp.
770–778.
[24] J. G. A. Barbedo, “Plant disease identification from individual lesions and spots using deep
learning,” Biosystems Engineering, vol. 180, pp. 96–107, April 2019.
[25] M. Tan and Q. V. Le, “Efficientnet: Rethinking model scaling for convolutional neural networks,” in Proceedings of the International Conference on Machine Learning, May 2019, pp.
6105–6114.
¨ Atila, M. U¸car, K. Akyol, and E. U¸car, “Plant leaf disease classification using efficientnet
[26] U.
deep learning model,” Ecological Informatics, vol. 61, p. 101182, March 2021.
[27] Y. Kawasaki, H. Uga, S. Kagiwada, and H. Iyatomi, “Basic study of automated diagnosis of
viral plant diseases using convolutional neural networks,” in Proceedings of the International
Symposium on Visual Computing, December 2015, pp. 638–645.
[28] E. Fujita, Y. Kawasaki, H. Uga, S. Kagiwada, and H. Iyatomi, “Basic investigation on a robust
and practical plant diagnostic system,” in Proceedings of the IEEE International Conference on
Machine Learning and Applications, December 2016, pp. 989–992.
[29] E. Fujita, H. Uga, S. Kagiwada, and H. Iyatomi, “A practical plant diagnosis system for field leaf
images and feature visualization,” International Journal of Engineering & Technology, vol. 7,
no. 4.11, pp. 49–54, October 2018.
[30] T. Hiroki, R. Kotani, S. Kagiwada, U. Hiroyuki, and H. Iyatomi, “Diagnosis of multiple cucumber infections with convolutional neural networks,” in Proceedings of the Applied Imagery
Pattern Recognition Workshop, October 2018, pp. 1–4.
73
[31] S. Sladojevic, M. Arsenovic, A. Anderla, D. Culibrk, and D. Stefanovic, “Deep neural networks
based recognition of plant diseases by leaf image classification,” Computational Intelligence and
Neuroscience, vol. 2016, p. 3289801, June 2016.
[32] A. Ramcharan, K. Baranowski, P. McCloskey, B. Ahmed, J. Legg, and D. P. Hughes, “Deep
learning for image-based cassava disease detection,” Frontiers in Plant Science, vol. 8, p. 1852,
October 2017.
[33] C. DeChant, T. Wiesner-Hanks, S. Chen, E. L. Stewart, J. Yosinski, M. A. Gore, R. J. Nelson,
and H. Lipson, “Automated identification of northern leaf blight-infected maize plants from field
imagery using deep learning,” Phytopathology, vol. 107, no. 11, pp. 1426–1432, November 2017.
[34] K. P. Ferentinos, “Deep learning models for plant disease detection and diagnosis,” Computers
and Electronics in Agriculture, vol. 145, pp. 311–318, February 2018.
[35] Y. Lu, S. Yi, N. Zeng, Y. Liu, and Y. Zhang, “Identification of rice diseases using deep convolutional neural networks,” Neurocomputing, vol. 267, pp. 378–384, December 2017.
[36] J. Chen, J. Chen, D. Zhang, Y. Sun, and Y. A. Nanehkaran, “Using deep transfer learning for
image-based plant disease identification,” Computers and Electronics in Agriculture, vol. 173,
p. 105393, June 2020.
[37] C. Szegedy, V. Vanhoucke, S. Ioffe, J. Shlens, and Z. Wojna, “Rethinking the inception architecture for computer vision,” in Proceedings of the IEEE Conference on Computer Vision and
Pattern Recognition, June 2016, pp. 2818–2826.
[38] A. Picon, M. Seitz, A. Alvarez-Gila, P. Mohnke, A. Ortiz-Barredo, and J. Echazarra, “Crop
conditional convolutional neural networks for massive multi-crop plant disease classification
over cell phone acquired images taken on real field conditions,” Computers and Electronics in
Agriculture, vol. 167, p. 105093, December 2019.
[39] A. F. Fuentes, S. Yoon, S. Kim, and D. S. Park, “A robust deep-learning-based detector for realtime tomato plant diseases and pests recognition,” Sensors, vol. 17, no. 9, p. 2022, September
2017.
[40] J. Lu, J. Hu, G. Zhao, F. Mei, and C. Zhang, “An in-field automatic wheat disease diagnosis
system,” Computers and Electronics in Agriculture, vol. 142, pp. 369–379, September 2017.
74
[41] Q. Wang, F. Qi, M. Sun, J. Qu, and J. Xue, “Identification of tomato disease types and detection
of infected areas based on deep convolutional neural networks and object detection techniques,”
Computational Intelligence and Neuroscience, vol. 2019, p. 9142753, December 2019.
[42] M. M. Ozguven and K. Adem, “Automatic detection and classification of leaf spot disease in
sugar beet using deep learning algorithms,” Physica A: Statistical Mechanics and its Applications, vol. 535, p. 122537, December 2019.
[43] P. V. Bhatt, S. Sarangi, and S. Pappula, “Detection of diseases and pests on images captured
in uncontrolled conditions from tea plantations,” in Proceedings of the Autonomous Air and
Ground Sensing Systems for Agricultural Optimization and Phenotyping IV, vol. 11008, May
2019, p. 1100808.
[44] P. Jiang, Y. Chen, B. Liu, D. He, and C. Liang, “Real-time detection of apple leaf diseases
using deep learning approach based on improved convolutional neural networks,” IEEE Access,
vol. 7, pp. 59 069–59 080, May 2019.
[45] X. Xie, Y. Ma, B. Liu, J. He, S. Li, and H. Wang, “A deep-learning-based real-time detector for
grape leaf diseases using improved convolutional neural networks,” Frontiers in plant science,
vol. 11, p. 751, June 2020.
[46] R. Girshick, J. Donahue, T. Darrell, and J. Malik, “Rich feature hierarchies for accurate object
detection and semantic segmentation,” in Proceedings of the IEEE Conference on Computer
Vision and Pattern Recognition, June 2014, pp. 580–587.
[47] R. Girshick, “Fast r-cnn,” in Proceedings of the IEEE International Conference on Computer
Vision, December 2015, pp. 1440–1448.
[48] S. Ren, K. He, R. Girshick, and J. Sun, “Faster r-cnn: towards real-time object detection with
region proposal networks,” IEEE Transactions on Pattern Analysis and Machine Intelligence,
vol. 39, no. 6, pp. 1137–1149, June 2016.
[49] J. Dai, Y. Li, K. He, and J. Sun, “R-fcn: Object detection via region-based fully convolutional
networks,” in Proceedings of the Advances in Neural Information Processing Systems, December
2016, pp. 379–387.
75
[50] J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, “You only look once: Unified, real-time
object detection,” in Proceedings of the IEEE Conference on Computer Vision and Pattern
Recognition, June 2016, pp. 779–788.
[51] W. Liu, D. Anguelov, D. Erhan, C. Szegedy, S. Reed, C.-Y. Fu, and A. C. Berg, “Ssd: Single shot
multibox detector,” in Proceedings of the European Conference on Computer Vision, October
2016, pp. 21–37.
[52] T. Y. Lin, P. Goyal, R. Girshick, K. He, and P. Doll´ar, “Focal loss for dense object detection,”
in Proceedings of the IEEE International Conference on Computer Vision, October 2017, pp.
2980–2988.
[53] J. Redmon and A. Farhadi, “Yolo9000: better, faster, stronger,” in Proceedings of the IEEE
Conference on Computer Vision and Pattern Recognition, June 2017, pp. 7263–7271.
[54] J. Redmon and A. Farhadi, “Yolov3: An incremental improvement,” arXiv:1804.02767, April
2018. [Online]. Available: https://arxiv.org/abs/1804.02767
[55] M. Tan, R. Pang, and Q. V. Le, “Efficientdet: Scalable and efficient object detection,” in
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, June 2020,
pp. 10 781–10 790.
[56] C.-Y. Wang,
A. Bochkovskiy,
and H.-Y. M. Liao,
“Scaled-yolov4:
Scaling cross
stage partial network,” arXiv:2011.08036, November 2020. [Online]. Available:
https:
//arxiv.org/abs/2011.08036
[57] K. He, G. Gkioxari, P. Doll´
ar, and R. Girshick, “Mask r-cnn,” in Proceedings of the IEEE
International Conference on Computer Vision, October 2017, pp. 2961–2969.
[58] A. K. Singh, B. Ganapathysubramanian, S. Sarkar, and A. Singh, “Deep learning for plant
stress phenotyping: trends and future perspectives,” Trends in Plant Science, vol. 23, no. 10,
pp. 883–898, October 2018.
[59] J. Liu and X. Wang, “Plant diseases and pests detection based on deep learning: a review,”
Plant Methods, vol. 17, no. 1, pp. 1–18, February 2021.
[60] I. Sa, Z. Ge, F. Dayoub, B. Upcroft, T. Perez, and C. McCool, “Deepfruits: A fruit detection
system using deep neural networks,” Sensors, vol. 16, no. 8, p. 1222, August 2016.
76
[61] K. Bresilla, G. D. Perulli, A. Boini, B. Morandi, L. Corelli Grappadelli, and L. Manfrini, “Singleshot convolution neural networks for real-time fruit detection within the tree,” Frontiers in Plant
Science, vol. 10, p. 611, May 2019.
[62] Y. Tian, G. Yang, Z. Wang, H. Wang, E. Li, and Z. Liang, “Apple detection during different
growth stages in orchards using the improved yolo-v3 model,” Computers and Electronics in
Agriculture, vol. 157, pp. 417–426, January 2019.
[63] Q. H. Cap, K. Suwa, E. Fujita, H. Uga, S. Kagiwada, and H. Iyatomi, “An end-to-end practical plant disease diagnosis system for wide-angle cucumber images,” International Journal of
Engineering & Technology, vol. 7, no. 4.11, pp. 106–111, October 2018.
[64] T. Saikawa, Q. H. Cap, S. Kagiwada, H. Uga, and H. Iyatomi, “Aop: An anti-overfitting pretreatment for practical image-based plant diagnosis,” in Proceedings of the IEEE International
Conference on Big Data Workshops, December 2019, pp. 5177–5182.
[65] K. Suwa, Q. H. Cap, R. Kotani, H. Uga, S. Kagiwada, and H. Iyatomi, “A comparable study:
Intrinsic difficulties of practical plant diagnosis from wide-angle images,” in Proceedings of the
IEEE International Conference on Big Data Workshops, December 2019, pp. 5195–5201.
[66] R. R. Selvaraju, M. Cogswell, A. Das, R. Vedantam, D. Parikh, and D. Batra, “Grad-cam:
Visual explanations from deep networks via gradient-based localization,” in Proceedings of the
IEEE International Conference on Computer Vision, October 2017, pp. 618–626.
[67] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair et al., “Generative
adversarial nets,” in Proceedings of the Advances in Neural Information Processing Systems,
December 2014, pp. 2672–2680.
[68] M. Arsenovic, M. Karanovic, S. Sladojevic, A. Anderla, and D. Stefanovic, “Solving current
limitations of deep learning based approaches for plant disease detection,” Symmetry, vol. 11,
no. 7, p. 939, July 2019.
[69] Y. Tian, G. Yang, Z. Wang, E. Li, and Z. Liang, “Detection of apple lesions in orchards based on
deep learning methods of cyclegan and yolov3-dense,” Journal of Sensors, vol. 2019, p. 7630926,
April 2019.
77
[70] H. Nazki, S. Yoon, A. Fuentes, and D. S. Park, “Unsupervised image translation using adversarial networks for improved plant disease recognition,” Computers and Electronics in Agriculture,
vol. 168, p. 105117, January 2020.
[71] H. Chen, M. Guan, and H. Li, “Arcyclegan: Improved cyclegan for style transferring of fruit
images,” IEEE Access, vol. 9, pp. 46 776–46 787, March 2021.
[72] Q. Wu, Y. Chen, and J. Meng, “Dcgan-based data augmentation for tomato leaf disease identification,” IEEE Access, vol. 8, pp. 98 716–98 728, May 2020.
[73] S. Kanno, S. Nagasawa, Q. H. Cap, S. Shibuya, H. Uga, S. Kagiwada, and H. Iyatomi, “Ppig:
Productive and pathogenic image generation for plant disease diagnosis,” in Proceedings of the
IEEE-EMBS Conference on Biomedical Engineering and Sciences, March 2021, pp. 554–559.
[74] N. Qian, “On the momentum term in gradient descent learning algorithms,” Neural Networks,
vol. 12, no. 1, pp. 145–151, January 1999.
[75] E. Real, A. Aggarwal, Y. Huang, and Q. V. Le, “Regularized evolution for image classifier
architecture search,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33,
July 2019, pp. 4780–4789.
[76] Y. Huang, Y. Cheng, A. Bapna, O. Firat, D. Chen, M. Chen, H. Lee, J. Ngiam, Q. V. Le, Y. Wu,
and Z. Chen, “Gpipe: Efficient training of giant neural networks using pipeline parallelism,”
in Proceedings of the Advances in Neural Information Processing Systems, December 2019, pp.
103–112.
[77] B. C. Tom and A. K. Katsaggelos, “Reconstruction of a high-resolution image by simultaneous registration, restoration, and interpolation of low-resolution images,” in Proceedings of the
International Conference on Image Processing, vol. 2, October 1995, pp. 539–542.
[78] R. R. Schultz and R. L. Stevenson, “Extraction of high-resolution frames from video sequences,”
IEEE Transactions on Image Processing, vol. 5, no. 6, pp. 996–1011, June 1996.
[79] A. J. Patti and Y. Altunbasak, “Artifact reduction for set theoretic super resolution image reconstruction with edge adaptive constraints and higher-order interpolants,” IEEE Transactions
on Image Processing, vol. 10, no. 1, pp. 179–186, January 2001.
[80] D. Hirao and H. Iyatomi, “Prototype of super-resolution camera array system,” in Proceedings
of the International Symposium on Visual Computing, December 2015, pp. 911–920.
78
[81] E. Quevedo, E. Delory, G. Callic´
o, F. Tobajas, and R. Sarmiento, “Underwater video enhancement using multi-camera super-resolution,” Optics Communications, vol. 404, pp. 94–102, December 2017.
[82] W. T. Freeman, T. R. Jones, and E. C. Pasztor, “Example-based super-resolution,” IEEE
Computer Graphics and Applications, vol. 22, no. 2, pp. 56–65, August 2002.
[83] T. Komatsu, Y. Ueda, and T. Saito, “Super-resolution decoding of jpeg-compressed image
data with the shrinkage in the redundant dct domain,” in Proceedings of the Picture Coding
Symposium, December 2010, pp. 114–117.
[84] C. Dong, C. C. Loy, K. He, and X. Tang, “Image super-resolution using deep convolutional
networks,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 38, no. 2, pp.
295–307, June 2015.
[85] C. Ledig, L. Theis, F. Husz´
ar, J. Caballero, A. Cunningham, A. Acosta, A. Aitken, A. Tejani,
J. Totz, Z. Wang, and W. Shi, “Photo-realistic single image super-resolution using a generative
adversarial network,” in Proceedings of the IEEE Conference on Computer Vision and Pattern
Recognition, July 2017, pp. 4681–4690.
[86] X. Wang, K. Yu, S. Wu, J. Gu, Y. Liu, C. Dong, Y. Qiao, and C. Change Loy, “Esrgan:
Enhanced super-resolution generative adversarial networks,” in Proceedings of the European
Conference on Computer Vision, September 2018, pp. 1–16.
[87] A. Jolicoeur-Martineau, “The relativistic discriminator: a key element missing from standard
gan,” in Proceedings of the International Conference on Learning Representations, May 2019,
pp. 1–26.
[88] S. B. Kasturiwala and S. Aladhake, “Adaptive image superresolution for agrobased application,”
in Proceedings of the International Conference on Industrial Instrumentation and Control, May
2015, pp. 650–655.
[89] K. Yamamoto, T. Togami, and N. Yamaguchi, “Super-resolution of plant disease images for the
acceleration of image-based phenotyping and vigor diagnosis in agriculture,” Sensors, vol. 17,
no. 11, p. 2557, November 2017.
79
[90] Q. H. Cap, H. Tani, H. Uga, S. Kagiwada, and H. Iyatomi, “Super-resolution for practical automated plant disease diagnosis system,” in Proceedings of the Annual Conference on Information
Sciences and Systems, March 2019, pp. 1–6.
[91] Q. Dai, X. Cheng, Y. Qiao, and Y. Zhang, “Crop leaf disease image super-resolution and
identification with dual attention and topology fusion generative adversarial network,” IEEE
Access, vol. 8, pp. 55 724–55 735, March 2020.
[92] A. Giachetti and N. Asuni, “Real-time artifact-free image upscaling,” IEEE Transactions on
Image Processing, vol. 20, no. 10, pp. 2760–2768, April 2011.
[93] J. Johnson, A. Alahi, and L. Fei-Fei, “Perceptual losses for real-time style transfer and superresolution,” in Proceedings of the European Conference on Computer Vision, October 2016, pp.
694–711.
[94] M. Heusel, H. Ramsauer, T. Unterthiner, B. Nessler, and S. Hochreiter, “Gans trained by a two
time-scale update rule converge to a local nash equilibrium,” in Proceedings of the Advances in
Neural Information Processing Systems, December 2017, pp. 6626–6637.
[95] A. L. Maas, A. Y. Hannun, and A. Y. Ng, “Rectifier nonlinearities improve neural network
acoustic models,” in Proceedings of the International Conference on Machine Learning, vol. 30,
no. 1, June 2013, p. 3.
[96] S. Ioffe and C. Szegedy, “Batch normalization: Accelerating deep network training by reducing
internal covariate shift,” in Proceedings of the International Conference on Machine Learning,
June 2015, pp. 448–456.
[97] D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” in Proceedings of the
International Conference on Learning Representations, May 2015, pp. 1–15.
[98] Y. Cui, M. Jia, T. Y. Lin, Y. Song, and S. Belongie, “Class-balanced loss based on effective
number of samples,” in Proceedings of the IEEE Conference on Computer Vision and Pattern
Recognition, June 2019, pp. 9268–9277.
[99] G. Toderici, D. Vincent, N. Johnston, S. Jin Hwang, D. Minnen, J. Shor, and M. Covell,
“Full resolution image compression with recurrent neural networks,” in Proceedings of the IEEE
Conference on Computer Vision and Pattern Recognition, July 2017, pp. 5306–5314.
80
[100] R. Zhang, P. Isola, A. A. Efros, E. Shechtman, and O. Wang, “The unreasonable effectiveness
of deep features as a perceptual metric,” in Proceedings of the IEEE Conference on Computer
Vision and Pattern Recognition, June 2018, pp. 586–595.
[101] A. F. Fuentes, S. Yoon, J. Lee, and D. S. Park, “High-performance deep neural network-based
tomato plant diseases and pests diagnosis system with refinement filter bank,” Frontiers in
Plant Science, vol. 9, p. 1162, August 2018.
[102] J.-Y. Zhu, T. Park, P. Isola, and A. A. Efros, “Unpaired image-to-image translation using
cycle-consistent adversarial networks,” in Proceedings of the IEEE International Conference on
Computer Vision, October 2017, pp. 2223–2232.
[103] Y. A. Mejjati, C. Richardt, J. Tompkin, D. Cosker, and K. I. Kim, “Unsupervised attentionguided image-to-image translation,” in Proceedings of the Advances in Neural Information Processing Systems, December 2018, pp. 3693–3703.
[104] X. Chen, C. Xu, X. Yang, and D. Tao, “Attention-gan for object transfiguration in wild images,”
in Proceedings of the European Conference on Computer Vision, September 2018, pp. 164–180.
[105] C. Yang, T. Kim, R. Wang, H. Peng, and C.-C. J. Kuo, “Show, attend, and translate: Unsupervised image translation with self-regularization and attention,” IEEE Transactions on Image
Processing, vol. 28, no. 10, pp. 4845–4856, May 2019.
[106] E. Shelhamer, J. Long, and T. Darrell, “Fully convolutional networks for semantic segmentation,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 39, no. 4, pp.
640–651, April 2017.
[107] O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image
segmentation,” in Proceedings of the International Conference on Medical Image Computing and
Computer-assisted Intervention, October 2015, pp. 234–241.
[108] H. Zhao, J. Shi, X. Qi, X. Wang, and J. Jia, “Pyramid scene parsing network,” in Proceedings of
the IEEE Conference on Computer Vision and Pattern Recognition, July 2017, pp. 2881–2890.
[109] B. Zhou, H. Zhao, X. Puig, T. Xiao, S. Fidler, A. Barriuso, and A. Torralba, “Semantic understanding of scenes through the ade20k dataset,” International Journal of Computer Vision, vol.
127, no. 3, pp. 302–321, March 2019.
81
[110] M. Oquab, L. Bottou, I. Laptev, and J. Sivic, “Is object localization for free?-weakly-supervised
learning with convolutional neural networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, June 2015, pp. 685–694.
[111] K. K. Singh and Y. J. Lee, “Hide-and-seek: Forcing a network to be meticulous for weaklysupervised object and action localization,” in Proceedings of the IEEE International Conference
on Computer Vision, October 2017, pp. 3544–3553.
[112] K. Li, Z. Wu, K.-C. Peng, J. Ernst, and Y. Fu, “Tell me where to look: Guided attention
inference network,” in Proceedings of the IEEE Conference on Computer Vision and Pattern
Recognition, June 2018, pp. 9215–9223.
[113] J. Lee, E. Kim, S. Lee, J. Lee, and S. Yoon, “Ficklenet: Weakly and semi-supervised semantic image segmentation using stochastic inference,” in Proceedings of the IEEE Conference on
Computer Vision and Pattern Recognition, June 2019, pp. 5267–5276.
[114] L. Perez and J. Wang, “The effectiveness of data augmentation in image classification
using deep learning,” arXiv:1712.04621, December 2017. [Online]. Available:
https:
//arxiv.org/abs/1712.04621
[115] C. Han, L. Rundo, R. Araki, Y. Furukawa, G. Mauri, H. Nakayama, and H. Hayashi, “Infinite
brain tumor images: Can gan-based data augmentation improve tumor detection on mr images?”
in Proceedings of the Meeting on Image Recognition and Understanding, August 2018, pp. 1–4.
[116] K. Shmelkov, C. Schmid, and K. Alahari, “How good is my gan?” in Proceedings of the European
Conference on Computer Vision, September 2018, pp. 213–229.
[117] Y. Ma, K. Liu, Z. Guan, X. Xu, X. Qian, and H. Bao, “Background augmentation generative
adversarial networks (bagans): Effective data generation based on gan-augmented 3d synthesizing,” Symmetry, vol. 10, no. 12, p. 734, December 2018.
[118] A. Radford, L. Metz, and S. Chintala, “Unsupervised representation learning with deep convolutional generative adversarial networks,” in Proceedings of the International Conference on
Learning Representations, May 2016, pp. 1–16.
[119] T. Karras, S. Laine, and T. Aila, “A style-based generator architecture for generative adversarial
networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition,
June 2019, pp. 4401–4410.
82
...