リケラボ論文検索は、全国の大学リポジトリにある学位論文・教授論文を一括検索できる論文検索サービスです。

リケラボ 全国の大学リポジトリにある学位論文・教授論文を一括検索するならリケラボ論文検索大学・研究所にある論文を検索できる

リケラボ 全国の大学リポジトリにある学位論文・教授論文を一括検索するならリケラボ論文検索大学・研究所にある論文を検索できる

大学・研究所にある論文を検索できる 「Deep Learning Approach for Practical Plant Disease Diagnosis」の論文概要。リケラボ論文検索は、全国の大学リポジトリにある学位論文・教授論文を一括検索できる論文検索サービスです。

コピーが完了しました

URLをコピーしました

論文の公開元へ論文の公開元へ
書き出し

Deep Learning Approach for Practical Plant Disease Diagnosis

HUU QUAN Cap 法政大学 DOI:info:doi/10.15002/00024532

2021.12.13

概要

With the breakthrough of deep learning techniques, many excellent applications for the automated diagnosis of plant disease have been proposed. However, there are several open issues for developing practical plant disease diagnosis systems in real cultivation. Firstly, most conventional methodologies only accept narrow range images, typically one or quite a limited number of targets are in their inputs. Applying these models to wide-angle images in large farms would be very time-consuming, since many targets (e.g., leaves) need to be diagnosed. In this work, we propose a two-stage system which has independent leaf detection and leaf diagnosis stages for wide-angle disease diagnosis. We show that our proposal attains a promising disease diagnostic performance that is more than six times higher than end-to-end systems (state-of-the-art detection methods like Faster R-CNN or SSD) with F1-score of 33.4 - 38.9% compared to 4.4 - 6.2% on an unseen target dataset.
Secondly, the lack of image resolution (i.e., diagnosing from low-quality input images such as low-resolution, blur, poor camera focus, etc.) could significantly reduce the diagnostic performance in practice. Also, high-resolution data is very difficult to obtain and are not always available in practice. Deep learning-based techniques, and particularly generative adversarial networks (GANs), can be applied to generate high-quality super-resolution images, but these methods often produce unexpected artifacts that can lower the diagnostic performance. In this paper, we propose a novel artifact-suppression super-resolution method that is specifically designed for diagnosing leaf disease, called LASSR. Our LASSR can detect and suppress artifacts to a considerable extent. Thus, generating much more pleasing, high-quality images from low-resolution inputs. Experiments show that training with data generated by our proposal significantly boosts the performance on an unseen test dataset by over 21% compared with the baseline.
Thirdly, collecting and labeling training disease data for these diagnosis systems requires solid biological knowledge and is very labor-intensive. Limited amount of disease training data leads to the fourth problem of model overfitting. The performance of disease diagnostic models are drastically decreased when used on test data sets from new environments. Meanwhile, we observe that healthy images are easier to collect. Based on this, we propose LeafGAN, a novel image-to-image translation system. LeafGAN generates countless diverse and high-quality diseased data via transformation from healthy images, as a data augmentation tool for improving the performance of plant disease diagnosis. Our model can transform only relevant areas from images with a variety of backgrounds, thus enriching the versatility of the training images. Experiments show that data augmentation with LeafGAN help to improve the generalization, boosting the diagnostic performance on unseen data by 7.4% from baseline.
In summary, we show that our approaches significantly improve the diagnostic performance under practical settings, confirming to be efficient and reliable methods for real cultivation scenarios.

この論文で使われている画像

参考文献

[1] R. N. Strange and P. R. Scott, “Plant disease: a threat to global food security,” Annual Review

of Phytopathology, vol. 43, pp. 83–116, August 2005.

[2] S. Savary, L. Willocquet, S. J. Pethybridge, P. Esker, N. McRoberts, and A. Nelson, “The global

burden of pathogens and pests on major food crops,” Nature Ecology & Evolution, vol. 3, no. 3,

pp. 430–439, March 2019.

[3] F. Qin, D. Liu, B. Sun, L. Ruan, Z. Ma, and H. Wang, “Identification of alfalfa leaf diseases

using image recognition technology,” PLoS One, vol. 11, no. 12, p. e0168274, December 2016.

[4] C. Cortes and V. Vapnik, “Support-vector networks,” Machine learning, vol. 20, no. 3, pp.

273–297, March 1995.

[5] L. Hallau, M. Neumann, B. Klatt, B. Kleinhenz, T. Klein, C. Kuhn, M. R¨ohrig, C. Bauckhage, K. Kersting, A.-K. Mahlein et al., “Automated identification of sugar beet diseases using

smartphones,” Plant Pathology, vol. 67, no. 2, pp. 399–410, February 2018.

[6] E. Mwebaze and G. Owomugisha, “Machine learning for plant disease incidence and severity

measurements from leaf images,” in Proceedings of the 15th IEEE International Conference on

Machine Learning and Applications, December 2016, pp. 158–163.

[7] E. Rublee, V. Rabaud, K. Konolige, and G. Bradski, “Orb: An efficient alternative to sift or

surf,” in Proceedings of the International Conference on Computer Vision, November 2011, pp.

2564–2571.

[8] Y. Es-saady, I. El Massi, M. El Yassa, D. Mammass, and A. Benazoun, “Automatic recognition

of plant leaves diseases based on serial combination of two svm classifiers,” in Proceedings of the

International Conference on Electrical and Information Technologies, May 2016, pp. 561–566.

[9] L. C. Ngugi, M. Abelwahab, and M. Abo-Zahhad, “Recent advances in image processing techniques for automated leaf pest and disease recognition-a review,” Information Processing in

Agriculture, vol. 8, no. 1, pp. 27–51, March 2021.

[10] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “Imagenet classification with deep convolutional

neural networks,” in Proceedings of the Advances in Neural Information Processing Systems,

vol. 25, December 2012, pp. 1097–1105.

71

[11] J. Deng, W. Dong, R. Socher, L. J. Li, K. Li, and L. Fei-Fei, “Imagenet: A large-scale hierarchical

image database,” in Proceedings of the IEEE Conference on Computer Vision and Pattern

Recognition, June 2009, pp. 248–255.

[12] B. Liu, Y. Zhang, D. He, and Y. Li, “Identification of apple leaf diseases based on deep convolutional neural networks,” Symmetry, vol. 10, no. 1, pp. 1–16, January 2018.

[13] C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke, and

A. Rabinovich, “Going deeper with convolutions,” in Proceedings of the IEEE Conference on

Computer Vision and Pattern Recognition, June 2015, pp. 1–9.

[14] S. P. Mohanty, D. P. Hughes, and M. Salath´e, “Using deep learning for image-based plant disease

detection,” Frontiers in Plant Science, vol. 7, p. 1419, September 2016.

[15] D. Hughes and M. Salath´e, “An open access repository of images on plant health to enable

the development of mobile disease diagnostics,” arXiv:1511.08060, November 2015. [Online].

Available: https://arxiv.org/abs/1511.08060

[16] G. Wang, Y. Sun, and J. Wang, “Automatic image-based plant disease severity estimation using

deep learning,” Computational Intelligence and Neuroscience, vol. 2017, p. 2917536, July 2017.

[17] K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” in Proceedings of the International Conference on Learning Representations, May 2015,

pp. 1–14.

[18] H. Durmu¸s, E. O. G¨

une¸s, and M. Kırcı, “Disease detection on the leaves of the tomato

plants by using deep learning,” in Proceedings of the 6th International Conference on AgroGeoinformatics, August 2017, pp. 1–5.

[19] F. N. Iandola, S. Han, M. W. Moskewicz, K. Ashraf, W. J. Dally, and K. Keutzer, “Squeezenet:

Alexnet-level accuracy with 50x fewer parameters and¡ 0.5 mb model size,” in Proceedings of

the International Conference on Learning Representations, April 2017, pp. 1–13.

[20] A. Elhassouny and F. Smarandache, “Smart mobile application to recognize tomato leaf diseases using convolutional neural networks,” in Proceedings of the International Conference of

Computer Science and Renewable Energies, July 2019, pp. 1–4.

72

[21] A. G. Howard, M. Zhu, B. Chen, D. Kalenichenko, W. Wang, T. Weyand, M. Andreetto, and

H. Adam, “Mobilenets: Efficient convolutional neural networks for mobile vision applications,”

arXiv:1704.04861, April 2017. [Online]. Available: https://arxiv.org/abs/1704.04861

[22] H. A. Atabay, “Deep residual learning for tomato plant leaf disease identification,” Journal of

Theoretical & Applied Information Technology, vol. 95, pp. 6800–6808, December 2017.

[23] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, June 2016, pp.

770–778.

[24] J. G. A. Barbedo, “Plant disease identification from individual lesions and spots using deep

learning,” Biosystems Engineering, vol. 180, pp. 96–107, April 2019.

[25] M. Tan and Q. V. Le, “Efficientnet: Rethinking model scaling for convolutional neural networks,” in Proceedings of the International Conference on Machine Learning, May 2019, pp.

6105–6114.

¨ Atila, M. U¸car, K. Akyol, and E. U¸car, “Plant leaf disease classification using efficientnet

[26] U.

deep learning model,” Ecological Informatics, vol. 61, p. 101182, March 2021.

[27] Y. Kawasaki, H. Uga, S. Kagiwada, and H. Iyatomi, “Basic study of automated diagnosis of

viral plant diseases using convolutional neural networks,” in Proceedings of the International

Symposium on Visual Computing, December 2015, pp. 638–645.

[28] E. Fujita, Y. Kawasaki, H. Uga, S. Kagiwada, and H. Iyatomi, “Basic investigation on a robust

and practical plant diagnostic system,” in Proceedings of the IEEE International Conference on

Machine Learning and Applications, December 2016, pp. 989–992.

[29] E. Fujita, H. Uga, S. Kagiwada, and H. Iyatomi, “A practical plant diagnosis system for field leaf

images and feature visualization,” International Journal of Engineering & Technology, vol. 7,

no. 4.11, pp. 49–54, October 2018.

[30] T. Hiroki, R. Kotani, S. Kagiwada, U. Hiroyuki, and H. Iyatomi, “Diagnosis of multiple cucumber infections with convolutional neural networks,” in Proceedings of the Applied Imagery

Pattern Recognition Workshop, October 2018, pp. 1–4.

73

[31] S. Sladojevic, M. Arsenovic, A. Anderla, D. Culibrk, and D. Stefanovic, “Deep neural networks

based recognition of plant diseases by leaf image classification,” Computational Intelligence and

Neuroscience, vol. 2016, p. 3289801, June 2016.

[32] A. Ramcharan, K. Baranowski, P. McCloskey, B. Ahmed, J. Legg, and D. P. Hughes, “Deep

learning for image-based cassava disease detection,” Frontiers in Plant Science, vol. 8, p. 1852,

October 2017.

[33] C. DeChant, T. Wiesner-Hanks, S. Chen, E. L. Stewart, J. Yosinski, M. A. Gore, R. J. Nelson,

and H. Lipson, “Automated identification of northern leaf blight-infected maize plants from field

imagery using deep learning,” Phytopathology, vol. 107, no. 11, pp. 1426–1432, November 2017.

[34] K. P. Ferentinos, “Deep learning models for plant disease detection and diagnosis,” Computers

and Electronics in Agriculture, vol. 145, pp. 311–318, February 2018.

[35] Y. Lu, S. Yi, N. Zeng, Y. Liu, and Y. Zhang, “Identification of rice diseases using deep convolutional neural networks,” Neurocomputing, vol. 267, pp. 378–384, December 2017.

[36] J. Chen, J. Chen, D. Zhang, Y. Sun, and Y. A. Nanehkaran, “Using deep transfer learning for

image-based plant disease identification,” Computers and Electronics in Agriculture, vol. 173,

p. 105393, June 2020.

[37] C. Szegedy, V. Vanhoucke, S. Ioffe, J. Shlens, and Z. Wojna, “Rethinking the inception architecture for computer vision,” in Proceedings of the IEEE Conference on Computer Vision and

Pattern Recognition, June 2016, pp. 2818–2826.

[38] A. Picon, M. Seitz, A. Alvarez-Gila, P. Mohnke, A. Ortiz-Barredo, and J. Echazarra, “Crop

conditional convolutional neural networks for massive multi-crop plant disease classification

over cell phone acquired images taken on real field conditions,” Computers and Electronics in

Agriculture, vol. 167, p. 105093, December 2019.

[39] A. F. Fuentes, S. Yoon, S. Kim, and D. S. Park, “A robust deep-learning-based detector for realtime tomato plant diseases and pests recognition,” Sensors, vol. 17, no. 9, p. 2022, September

2017.

[40] J. Lu, J. Hu, G. Zhao, F. Mei, and C. Zhang, “An in-field automatic wheat disease diagnosis

system,” Computers and Electronics in Agriculture, vol. 142, pp. 369–379, September 2017.

74

[41] Q. Wang, F. Qi, M. Sun, J. Qu, and J. Xue, “Identification of tomato disease types and detection

of infected areas based on deep convolutional neural networks and object detection techniques,”

Computational Intelligence and Neuroscience, vol. 2019, p. 9142753, December 2019.

[42] M. M. Ozguven and K. Adem, “Automatic detection and classification of leaf spot disease in

sugar beet using deep learning algorithms,” Physica A: Statistical Mechanics and its Applications, vol. 535, p. 122537, December 2019.

[43] P. V. Bhatt, S. Sarangi, and S. Pappula, “Detection of diseases and pests on images captured

in uncontrolled conditions from tea plantations,” in Proceedings of the Autonomous Air and

Ground Sensing Systems for Agricultural Optimization and Phenotyping IV, vol. 11008, May

2019, p. 1100808.

[44] P. Jiang, Y. Chen, B. Liu, D. He, and C. Liang, “Real-time detection of apple leaf diseases

using deep learning approach based on improved convolutional neural networks,” IEEE Access,

vol. 7, pp. 59 069–59 080, May 2019.

[45] X. Xie, Y. Ma, B. Liu, J. He, S. Li, and H. Wang, “A deep-learning-based real-time detector for

grape leaf diseases using improved convolutional neural networks,” Frontiers in plant science,

vol. 11, p. 751, June 2020.

[46] R. Girshick, J. Donahue, T. Darrell, and J. Malik, “Rich feature hierarchies for accurate object

detection and semantic segmentation,” in Proceedings of the IEEE Conference on Computer

Vision and Pattern Recognition, June 2014, pp. 580–587.

[47] R. Girshick, “Fast r-cnn,” in Proceedings of the IEEE International Conference on Computer

Vision, December 2015, pp. 1440–1448.

[48] S. Ren, K. He, R. Girshick, and J. Sun, “Faster r-cnn: towards real-time object detection with

region proposal networks,” IEEE Transactions on Pattern Analysis and Machine Intelligence,

vol. 39, no. 6, pp. 1137–1149, June 2016.

[49] J. Dai, Y. Li, K. He, and J. Sun, “R-fcn: Object detection via region-based fully convolutional

networks,” in Proceedings of the Advances in Neural Information Processing Systems, December

2016, pp. 379–387.

75

[50] J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, “You only look once: Unified, real-time

object detection,” in Proceedings of the IEEE Conference on Computer Vision and Pattern

Recognition, June 2016, pp. 779–788.

[51] W. Liu, D. Anguelov, D. Erhan, C. Szegedy, S. Reed, C.-Y. Fu, and A. C. Berg, “Ssd: Single shot

multibox detector,” in Proceedings of the European Conference on Computer Vision, October

2016, pp. 21–37.

[52] T. Y. Lin, P. Goyal, R. Girshick, K. He, and P. Doll´ar, “Focal loss for dense object detection,”

in Proceedings of the IEEE International Conference on Computer Vision, October 2017, pp.

2980–2988.

[53] J. Redmon and A. Farhadi, “Yolo9000: better, faster, stronger,” in Proceedings of the IEEE

Conference on Computer Vision and Pattern Recognition, June 2017, pp. 7263–7271.

[54] J. Redmon and A. Farhadi, “Yolov3: An incremental improvement,” arXiv:1804.02767, April

2018. [Online]. Available: https://arxiv.org/abs/1804.02767

[55] M. Tan, R. Pang, and Q. V. Le, “Efficientdet: Scalable and efficient object detection,” in

Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, June 2020,

pp. 10 781–10 790.

[56] C.-Y. Wang,

A. Bochkovskiy,

and H.-Y. M. Liao,

“Scaled-yolov4:

Scaling cross

stage partial network,” arXiv:2011.08036, November 2020. [Online]. Available:

https:

//arxiv.org/abs/2011.08036

[57] K. He, G. Gkioxari, P. Doll´

ar, and R. Girshick, “Mask r-cnn,” in Proceedings of the IEEE

International Conference on Computer Vision, October 2017, pp. 2961–2969.

[58] A. K. Singh, B. Ganapathysubramanian, S. Sarkar, and A. Singh, “Deep learning for plant

stress phenotyping: trends and future perspectives,” Trends in Plant Science, vol. 23, no. 10,

pp. 883–898, October 2018.

[59] J. Liu and X. Wang, “Plant diseases and pests detection based on deep learning: a review,”

Plant Methods, vol. 17, no. 1, pp. 1–18, February 2021.

[60] I. Sa, Z. Ge, F. Dayoub, B. Upcroft, T. Perez, and C. McCool, “Deepfruits: A fruit detection

system using deep neural networks,” Sensors, vol. 16, no. 8, p. 1222, August 2016.

76

[61] K. Bresilla, G. D. Perulli, A. Boini, B. Morandi, L. Corelli Grappadelli, and L. Manfrini, “Singleshot convolution neural networks for real-time fruit detection within the tree,” Frontiers in Plant

Science, vol. 10, p. 611, May 2019.

[62] Y. Tian, G. Yang, Z. Wang, H. Wang, E. Li, and Z. Liang, “Apple detection during different

growth stages in orchards using the improved yolo-v3 model,” Computers and Electronics in

Agriculture, vol. 157, pp. 417–426, January 2019.

[63] Q. H. Cap, K. Suwa, E. Fujita, H. Uga, S. Kagiwada, and H. Iyatomi, “An end-to-end practical plant disease diagnosis system for wide-angle cucumber images,” International Journal of

Engineering & Technology, vol. 7, no. 4.11, pp. 106–111, October 2018.

[64] T. Saikawa, Q. H. Cap, S. Kagiwada, H. Uga, and H. Iyatomi, “Aop: An anti-overfitting pretreatment for practical image-based plant diagnosis,” in Proceedings of the IEEE International

Conference on Big Data Workshops, December 2019, pp. 5177–5182.

[65] K. Suwa, Q. H. Cap, R. Kotani, H. Uga, S. Kagiwada, and H. Iyatomi, “A comparable study:

Intrinsic difficulties of practical plant diagnosis from wide-angle images,” in Proceedings of the

IEEE International Conference on Big Data Workshops, December 2019, pp. 5195–5201.

[66] R. R. Selvaraju, M. Cogswell, A. Das, R. Vedantam, D. Parikh, and D. Batra, “Grad-cam:

Visual explanations from deep networks via gradient-based localization,” in Proceedings of the

IEEE International Conference on Computer Vision, October 2017, pp. 618–626.

[67] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair et al., “Generative

adversarial nets,” in Proceedings of the Advances in Neural Information Processing Systems,

December 2014, pp. 2672–2680.

[68] M. Arsenovic, M. Karanovic, S. Sladojevic, A. Anderla, and D. Stefanovic, “Solving current

limitations of deep learning based approaches for plant disease detection,” Symmetry, vol. 11,

no. 7, p. 939, July 2019.

[69] Y. Tian, G. Yang, Z. Wang, E. Li, and Z. Liang, “Detection of apple lesions in orchards based on

deep learning methods of cyclegan and yolov3-dense,” Journal of Sensors, vol. 2019, p. 7630926,

April 2019.

77

[70] H. Nazki, S. Yoon, A. Fuentes, and D. S. Park, “Unsupervised image translation using adversarial networks for improved plant disease recognition,” Computers and Electronics in Agriculture,

vol. 168, p. 105117, January 2020.

[71] H. Chen, M. Guan, and H. Li, “Arcyclegan: Improved cyclegan for style transferring of fruit

images,” IEEE Access, vol. 9, pp. 46 776–46 787, March 2021.

[72] Q. Wu, Y. Chen, and J. Meng, “Dcgan-based data augmentation for tomato leaf disease identification,” IEEE Access, vol. 8, pp. 98 716–98 728, May 2020.

[73] S. Kanno, S. Nagasawa, Q. H. Cap, S. Shibuya, H. Uga, S. Kagiwada, and H. Iyatomi, “Ppig:

Productive and pathogenic image generation for plant disease diagnosis,” in Proceedings of the

IEEE-EMBS Conference on Biomedical Engineering and Sciences, March 2021, pp. 554–559.

[74] N. Qian, “On the momentum term in gradient descent learning algorithms,” Neural Networks,

vol. 12, no. 1, pp. 145–151, January 1999.

[75] E. Real, A. Aggarwal, Y. Huang, and Q. V. Le, “Regularized evolution for image classifier

architecture search,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33,

July 2019, pp. 4780–4789.

[76] Y. Huang, Y. Cheng, A. Bapna, O. Firat, D. Chen, M. Chen, H. Lee, J. Ngiam, Q. V. Le, Y. Wu,

and Z. Chen, “Gpipe: Efficient training of giant neural networks using pipeline parallelism,”

in Proceedings of the Advances in Neural Information Processing Systems, December 2019, pp.

103–112.

[77] B. C. Tom and A. K. Katsaggelos, “Reconstruction of a high-resolution image by simultaneous registration, restoration, and interpolation of low-resolution images,” in Proceedings of the

International Conference on Image Processing, vol. 2, October 1995, pp. 539–542.

[78] R. R. Schultz and R. L. Stevenson, “Extraction of high-resolution frames from video sequences,”

IEEE Transactions on Image Processing, vol. 5, no. 6, pp. 996–1011, June 1996.

[79] A. J. Patti and Y. Altunbasak, “Artifact reduction for set theoretic super resolution image reconstruction with edge adaptive constraints and higher-order interpolants,” IEEE Transactions

on Image Processing, vol. 10, no. 1, pp. 179–186, January 2001.

[80] D. Hirao and H. Iyatomi, “Prototype of super-resolution camera array system,” in Proceedings

of the International Symposium on Visual Computing, December 2015, pp. 911–920.

78

[81] E. Quevedo, E. Delory, G. Callic´

o, F. Tobajas, and R. Sarmiento, “Underwater video enhancement using multi-camera super-resolution,” Optics Communications, vol. 404, pp. 94–102, December 2017.

[82] W. T. Freeman, T. R. Jones, and E. C. Pasztor, “Example-based super-resolution,” IEEE

Computer Graphics and Applications, vol. 22, no. 2, pp. 56–65, August 2002.

[83] T. Komatsu, Y. Ueda, and T. Saito, “Super-resolution decoding of jpeg-compressed image

data with the shrinkage in the redundant dct domain,” in Proceedings of the Picture Coding

Symposium, December 2010, pp. 114–117.

[84] C. Dong, C. C. Loy, K. He, and X. Tang, “Image super-resolution using deep convolutional

networks,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 38, no. 2, pp.

295–307, June 2015.

[85] C. Ledig, L. Theis, F. Husz´

ar, J. Caballero, A. Cunningham, A. Acosta, A. Aitken, A. Tejani,

J. Totz, Z. Wang, and W. Shi, “Photo-realistic single image super-resolution using a generative

adversarial network,” in Proceedings of the IEEE Conference on Computer Vision and Pattern

Recognition, July 2017, pp. 4681–4690.

[86] X. Wang, K. Yu, S. Wu, J. Gu, Y. Liu, C. Dong, Y. Qiao, and C. Change Loy, “Esrgan:

Enhanced super-resolution generative adversarial networks,” in Proceedings of the European

Conference on Computer Vision, September 2018, pp. 1–16.

[87] A. Jolicoeur-Martineau, “The relativistic discriminator: a key element missing from standard

gan,” in Proceedings of the International Conference on Learning Representations, May 2019,

pp. 1–26.

[88] S. B. Kasturiwala and S. Aladhake, “Adaptive image superresolution for agrobased application,”

in Proceedings of the International Conference on Industrial Instrumentation and Control, May

2015, pp. 650–655.

[89] K. Yamamoto, T. Togami, and N. Yamaguchi, “Super-resolution of plant disease images for the

acceleration of image-based phenotyping and vigor diagnosis in agriculture,” Sensors, vol. 17,

no. 11, p. 2557, November 2017.

79

[90] Q. H. Cap, H. Tani, H. Uga, S. Kagiwada, and H. Iyatomi, “Super-resolution for practical automated plant disease diagnosis system,” in Proceedings of the Annual Conference on Information

Sciences and Systems, March 2019, pp. 1–6.

[91] Q. Dai, X. Cheng, Y. Qiao, and Y. Zhang, “Crop leaf disease image super-resolution and

identification with dual attention and topology fusion generative adversarial network,” IEEE

Access, vol. 8, pp. 55 724–55 735, March 2020.

[92] A. Giachetti and N. Asuni, “Real-time artifact-free image upscaling,” IEEE Transactions on

Image Processing, vol. 20, no. 10, pp. 2760–2768, April 2011.

[93] J. Johnson, A. Alahi, and L. Fei-Fei, “Perceptual losses for real-time style transfer and superresolution,” in Proceedings of the European Conference on Computer Vision, October 2016, pp.

694–711.

[94] M. Heusel, H. Ramsauer, T. Unterthiner, B. Nessler, and S. Hochreiter, “Gans trained by a two

time-scale update rule converge to a local nash equilibrium,” in Proceedings of the Advances in

Neural Information Processing Systems, December 2017, pp. 6626–6637.

[95] A. L. Maas, A. Y. Hannun, and A. Y. Ng, “Rectifier nonlinearities improve neural network

acoustic models,” in Proceedings of the International Conference on Machine Learning, vol. 30,

no. 1, June 2013, p. 3.

[96] S. Ioffe and C. Szegedy, “Batch normalization: Accelerating deep network training by reducing

internal covariate shift,” in Proceedings of the International Conference on Machine Learning,

June 2015, pp. 448–456.

[97] D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” in Proceedings of the

International Conference on Learning Representations, May 2015, pp. 1–15.

[98] Y. Cui, M. Jia, T. Y. Lin, Y. Song, and S. Belongie, “Class-balanced loss based on effective

number of samples,” in Proceedings of the IEEE Conference on Computer Vision and Pattern

Recognition, June 2019, pp. 9268–9277.

[99] G. Toderici, D. Vincent, N. Johnston, S. Jin Hwang, D. Minnen, J. Shor, and M. Covell,

“Full resolution image compression with recurrent neural networks,” in Proceedings of the IEEE

Conference on Computer Vision and Pattern Recognition, July 2017, pp. 5306–5314.

80

[100] R. Zhang, P. Isola, A. A. Efros, E. Shechtman, and O. Wang, “The unreasonable effectiveness

of deep features as a perceptual metric,” in Proceedings of the IEEE Conference on Computer

Vision and Pattern Recognition, June 2018, pp. 586–595.

[101] A. F. Fuentes, S. Yoon, J. Lee, and D. S. Park, “High-performance deep neural network-based

tomato plant diseases and pests diagnosis system with refinement filter bank,” Frontiers in

Plant Science, vol. 9, p. 1162, August 2018.

[102] J.-Y. Zhu, T. Park, P. Isola, and A. A. Efros, “Unpaired image-to-image translation using

cycle-consistent adversarial networks,” in Proceedings of the IEEE International Conference on

Computer Vision, October 2017, pp. 2223–2232.

[103] Y. A. Mejjati, C. Richardt, J. Tompkin, D. Cosker, and K. I. Kim, “Unsupervised attentionguided image-to-image translation,” in Proceedings of the Advances in Neural Information Processing Systems, December 2018, pp. 3693–3703.

[104] X. Chen, C. Xu, X. Yang, and D. Tao, “Attention-gan for object transfiguration in wild images,”

in Proceedings of the European Conference on Computer Vision, September 2018, pp. 164–180.

[105] C. Yang, T. Kim, R. Wang, H. Peng, and C.-C. J. Kuo, “Show, attend, and translate: Unsupervised image translation with self-regularization and attention,” IEEE Transactions on Image

Processing, vol. 28, no. 10, pp. 4845–4856, May 2019.

[106] E. Shelhamer, J. Long, and T. Darrell, “Fully convolutional networks for semantic segmentation,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 39, no. 4, pp.

640–651, April 2017.

[107] O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image

segmentation,” in Proceedings of the International Conference on Medical Image Computing and

Computer-assisted Intervention, October 2015, pp. 234–241.

[108] H. Zhao, J. Shi, X. Qi, X. Wang, and J. Jia, “Pyramid scene parsing network,” in Proceedings of

the IEEE Conference on Computer Vision and Pattern Recognition, July 2017, pp. 2881–2890.

[109] B. Zhou, H. Zhao, X. Puig, T. Xiao, S. Fidler, A. Barriuso, and A. Torralba, “Semantic understanding of scenes through the ade20k dataset,” International Journal of Computer Vision, vol.

127, no. 3, pp. 302–321, March 2019.

81

[110] M. Oquab, L. Bottou, I. Laptev, and J. Sivic, “Is object localization for free?-weakly-supervised

learning with convolutional neural networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, June 2015, pp. 685–694.

[111] K. K. Singh and Y. J. Lee, “Hide-and-seek: Forcing a network to be meticulous for weaklysupervised object and action localization,” in Proceedings of the IEEE International Conference

on Computer Vision, October 2017, pp. 3544–3553.

[112] K. Li, Z. Wu, K.-C. Peng, J. Ernst, and Y. Fu, “Tell me where to look: Guided attention

inference network,” in Proceedings of the IEEE Conference on Computer Vision and Pattern

Recognition, June 2018, pp. 9215–9223.

[113] J. Lee, E. Kim, S. Lee, J. Lee, and S. Yoon, “Ficklenet: Weakly and semi-supervised semantic image segmentation using stochastic inference,” in Proceedings of the IEEE Conference on

Computer Vision and Pattern Recognition, June 2019, pp. 5267–5276.

[114] L. Perez and J. Wang, “The effectiveness of data augmentation in image classification

using deep learning,” arXiv:1712.04621, December 2017. [Online]. Available:

https:

//arxiv.org/abs/1712.04621

[115] C. Han, L. Rundo, R. Araki, Y. Furukawa, G. Mauri, H. Nakayama, and H. Hayashi, “Infinite

brain tumor images: Can gan-based data augmentation improve tumor detection on mr images?”

in Proceedings of the Meeting on Image Recognition and Understanding, August 2018, pp. 1–4.

[116] K. Shmelkov, C. Schmid, and K. Alahari, “How good is my gan?” in Proceedings of the European

Conference on Computer Vision, September 2018, pp. 213–229.

[117] Y. Ma, K. Liu, Z. Guan, X. Xu, X. Qian, and H. Bao, “Background augmentation generative

adversarial networks (bagans): Effective data generation based on gan-augmented 3d synthesizing,” Symmetry, vol. 10, no. 12, p. 734, December 2018.

[118] A. Radford, L. Metz, and S. Chintala, “Unsupervised representation learning with deep convolutional generative adversarial networks,” in Proceedings of the International Conference on

Learning Representations, May 2016, pp. 1–16.

[119] T. Karras, S. Laine, and T. Aila, “A style-based generator architecture for generative adversarial

networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition,

June 2019, pp. 4401–4410.

82

...

参考文献をもっと見る