Fully Connected to Fully Convolutional: Road to Yesterday

Document Type : Persian Original Article

Author

Faculty of Mathematics and Computer Science

Abstract

In the last decade, several convolutional networks have been developed for the semantic segmentation, which have shown excellent performance in recognizing and labeling objects in images. Most of these networks involve large-scale architectures that can detect tens or hundreds of predefined classes. With the exception of fully convolutional networks, most applications use architectures that, after convolutional layers, use a common classifier to classify the extracted features. In this paper, the method of converting a network, which as classifier, has two flatten and dense layers (fully connected), to a fully convolutional network is described. The main advantage of this method is the ability to work on inputs of variable size and produce an output map instead of a number, which is the advantage of fully convolutional networks. Newer models of the Deep Learning area generally use training images in which areas of interest are determined by masks; but in the proposed method only labeled images are given to the network. The details of the proposed method are expressed in the form of a new problem of classification of boards with calligraphy of Shekasteh-Nastaliq and Suls, and classification of apple leaf diseases (as two-class problems) and the problem of identifying hand written Persian digits. For this purpose, first a convolutional network with the last fully connected layer is designed and trained for square images. Then a new fully convolutional model is defined based on the previous model and the weights of the previous model are fed to the new model. The only difference between the two models is in the last layer, but the new model will be able to work on input images of any size. Experimental results show the efficiency of the proposed approach.

Keywords


  1. S. A. Zaidi, M. S. Ansari, A. Aslam, N. Kanwal, M. Asghar, and B. Lee, "A survey of modern deep learning based object detection models," Digital Signal Processing, vol. 126, p. 103514, 2022.
  2. Schellenberg et al., "Semantic segmentation of multispectral photoacoustic images using deep learning," Photoacoustics, vol. 26, p. 100341, 2022.
  3. Shelhamer, J. Long, and T. Darrell, “Fully convolutional networks for semantic segmentation,” IEEE Trans. Pattern Anal. Mach. Intell., vol.39, p.640–651, Apr. 2017.
  4. Girshick, J. Donahue, T. Darrell, and J. Malik, “Rich feature hierarchies for accurate object detection and semantic segmentation,” in 2014 IEEE Conference on Computer Vision and Pattern Recognition, pp.580–587, 2014.
  5. Girshick, “Fast R-CNN,” in Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), ICCV ’15, (USA), p.1440–1448, IEEE Computer Society, 2015.
  6. Ren, K. He, R. Girshick, and J. Sun, “Faster R-CNN: Towards real-time object detection with region proposal networks,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol.39, no.6, pp.1137–1149, 2017.
  7. He, G. Gkioxari, P. Dollár, and R. Girshick, “Mask RCNN,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol.42, no.2, pp.386–397, 2020.
  8. Redmon, S. Divvala, R. Girshick, and A. Farhadi, “You only look once: Unified, real-time object detection,” in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp.779–788, 2016.
  9. Redmon and A. Farhadi, “Yolo9000: Better, faster, stronger,” CVPR, 2017.
  10. Badrinarayanan, A. Kendall, and R. Cipolla, “SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol.39, no.12, pp.2481–2495, 2017.
  11. Silberman, D. Hoiem, P. Kohli, and R. Fergus, “Indoor segmentation and support inference from RGBD images,” in Proceedings of the 12th European Conference on Computer Vision - Volume Part V, ECCV’12, (Berlin, Heidelberg), p.746–760, Springer-Verlag, 2012.
  12. Everingham, L. Van Gool, C. K. I. Williams, J. Winn, and A. Zisserman, “The Pascal Visual Object Classes (VOC) challenge,” International Journal of Computer Vision, vol.88, pp.303–338, June 2010.
  13. Mottaghi, X. Chen, X. Liu, N.-G. Cho, S.-W. Lee, S. Fidler, R. Urtasun, and A. Yuille, “The role of context for object detection and semantic segmentation in the wild,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2014.
  14. -Y. Lin, M. Maire, S. Belongie, J. Hays, P. Perona, D. Ramanan, P. Dollár, and C. L. Zitnick, “Microsoft coco: Common objects in context,” in Computer Vision – ECCV 2014 (D. Fleet, T. Pajdla, B. Schiele, and T. Tuytelaars, eds. ), (Cham), pp.740–755, Springer International Publishing, 2014.
  15. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” in International Conference on Learning Representations, 2015.
  16. Krizhevsky, I. Sutskever, and G. E. Hinton, “Imagenet classification with deep convolutional neural networks,” in Advances in Neural Information Processing Systems 25 (F. Pereira, C. J. C. Burges, L. Bottou, and K. Q. Weinberger, eds.), pp.1097–1105, Curran Associates, Inc., 2012.
  17. محمد صبری، محمد شهرام معین، و فربد رزازی،”ارائه‌ی روش ترتیبی پویا بر اساس یادگیری عمیق به منظور بهبود کارایی سیستم های تطبیق بیومتری مبتنی بر کارت هوشمند“، نشریه مهندس برق و مهندسی کامپیوتر، جلد ۱۸، شماره ۱-ب، صفحات ۲۹-۴۱. ۱۳۹۹.
  18. Li, M. Kan, S. Shan, and X. Chen, “Weakly supervised object detection with segmentation collaboration,” in Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), October 2019.
  19. Sheykhivand, S. Meshgini, and Z. Mousavi, “Automatic detection of various epileptic seizures from eeg signal using deeplearningnetworks,” Computational Intelligence in Electrical Engineering, vol.11, no.3, pp.1–12, 2020.
  20. Jiang, Y. Chen, B. Liu, D. He, and C. Liang, “Real-time detection of apple leaf diseases using deep learning approach based on improved convolutional neural networks,” IEEE Access, vol.7, pp.59069–59080, 2019.
  21. Chao, G. Sun, H. Zhao, M. Li, and D. He, “Identification of apple tree leaf diseases based on deep learning models,” Symmetry, vol.12, p.17, 06 2020.
  22. B. Tahir, M. A. Khan, K. Javed, S. Kadry, Y.-D. Zhang, T. Akram, and M. Nazir, “Recognition of apple leaf diseases using deep learning and variances-controlled features reduction,” Microprocessors and Microsystems, p.104027, 2021.
  23. Khosravi and E. Kabir, “Introducing a very large dataset of handwritten farsi digits and a study on their varieties,” Pattern Recognit. Lett., vol.28, pp.1133–1141, 2007.