Username   Password       Forgot your password?  Forgot your username? 

 

Classification of Remote Sensing Images based on Distributed Convolutional Neural Network Model

Volume 15, Number 6, June 2019, pp. 1508-1517
DOI: 10.23940/ijpe.19.06.p2.15081517

Guanyu Chena,b, Zhihua Caia,b, and Xiang Lia,b

aSchool of Computer Science, China University of Geosciences, Wuhan, 430074, China
bHubei Key Laboratory of Intelligent Geo-Information Processing, China University of Geosciences, Wuhan, 430074, China

 

(Submitted on December 12, 2018; Revised on January 15, 2019; Accepted on February 10, 2019)

Abstract:

With the network model architecture of Google Inception, research is conducted on issues such as the structural design of the model, data preprocessing, tuning of training parameters, computing clusters in a distributed environment, and multi-machine parallel training. According to the performances of different deep neural network models on different data sets, the Google Inception V3 depth network model is used as the prototype to conduct the tuning of training parameters, and the classification of remote sensing images is then realized with this model in the single-machine environment. Furthermore, due to the effectiveness of distributed systems for very large data sets and compute-intensive applications, a data parallel training scheme based on the distributed platform is designed for the convolution neural network model with more complex data form, larger quantity of parameters, and more network levels, after studying the mainstream designs of the distributed machine learning and analyzing the training methods and steps of the convolutional neural network model in a multi-machine environment. It greatly improves the training time of the model, and then the classification of remote sensing images under distributed clusters is realized.

 

References: 18

  1. D. Li, “Deep Learning: From Speech Recognition to Language and Multimodal Processing,” Apsipa Transactions on Signal & Information Processing, Vol. 5, No. 1, pp. 1-15, January 2016
  2. A. Sami, N. Kothari, J. Lee, P. Natsev, G. Toderici, B. Varadarajan, et al., “Youtube-8m: A Large-scale Video Classification Benchmark,” arXiv preprint arXiv: 1609.08675, September 2016
  3. G. Y. Cai and B. B. Xia, “Multimedia Sentiment Analysis based on Convolutional Neural Network,” Journal of Computer Applications, Vol. 36, No. 2, pp. 428-431, February 2016
  4. Q. Zhang, “Research on Convolutional Neural Network in Vehicle Logo Recognition Technology and Processing Strategy with Small Sample,” thesis, Anhui University, June 2016
  5. Y. Zhang and H. J. Yuan, “A Human Action Recognition Method based on 3D Convolution Neural Network,” Software Guide, Vol. 16, No. 11, pp. 9-11, November 2017
  6. H. F. Ma, X. B. Zhao, and X. C. Zou, “Research on Convolutional Neural Network Algorithm based on MapReduce,” Chinese Journal of Stereology and Image Analysis, Vol. 20, No. 4, pp. 339-346, December 2017
  7. A. Alhamali, N. Salha, R. Morcel, M. Ezzeddine, O. Hamdan, H. Akkary, et al., “FPGA-Accelerated Hadoop Cluster for Deep Learning Computations,” in Proceedings of 2015 IEEE International Conference on Data Mining Workshop (ICDMW), pp. 565-579, New York, USA, November 2015
  8. K. Y. Cheng, Y. Z. Zhan, and M. Qi, “AL-DDCNN: A Distributed Crossing Semantic Gap Learning for Person Re-identification,” Concurrency & Computation Practice & Experience, Vol. 29, No. 3, pp. 115-125, February 2017
  9. C. Szegedy, W. Liu, Y. Q. Jia, P. Sermanet, S. Reed, D. Anguelov, et al., “Going Deeper with Convolutions,” in Proceedings of 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1-9, Boston, USA, June 2015
  10. S. Ioffe and C. Szegedy, “Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift,” in Proceedings of International Conference on Machine Learning, pp. 448-456, Lille, France, July 2015
  11. C. Szegedy, V. Vanhoucke, S. Ioffe, J. Shlens, and Z. Wojna, “Rethinking the Inception Architecture for Computer Vision,” in Proceedings of 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2818-2826, Las Vegas, USA, July 2015
  12. C. Szegedy, S. Ioffe, and V. Vanhoucke, “Inception-V4, Inception-ResNet and the Impact of Residual Connections on Learning,” arXiv preprint arXiv: 1602.07261, February 2016
  13. J. Dean, G. S. Corrado, R. Monga, and K. Chen, “Large Scale Distributed Deep Networks,” in Proceedings of 2012 International Conference on Neural Information Processing Systems, pp. 1223-1231, Lake Tahoe, USA, December 2015
  14. J. M. Chen, X. H. Pan, R. Monga, S. Bengio, and R. Jozefowicz, “Revisiting Distributed Synchronous SGD,” arXiv preprint arXiv: 1604. 00981, April 2016
  15. J. E. Gonzalez, P. Bailis, M. I. Jordan, M. J. Franklin, J. M. Hellerstein, A. Ghodsi, et al., “Asynchronous Complex Analytics in a Distributed Dataflow Architecture,” arXiv preprint arXiv: 1510. 07092, October 2015
  16. D. Das, S. Avancha, D. Mudigere, K. Vaidynathan, S. Sridharan, D. Kalamkar, et al., “Distributed Deep Learning using Synchronous Stochastic Gradient Descent,” arXiv preprint arXiv: 1602.06709, February 2016
  17. J. W. Shi, Y. J. Qiu, U. F. Minhas, L. Jiao, C. Wang, B. Reinwald, et al., “Clash of the Titans: MapReduce vs. Spark for Large Scale Data Analytics,” Proceedings of the Vldb Endowment, Vol. 8, No. 13, pp. 2110-2121, January 2015
  18. G. Cheng, J. W. Han, and X. Q. Lu, “Remote Sensing Image Scene Classification: Benchmark and State of the Art,” Proceedings of the IEEE, Vol. 105, No. 10, pp. 1865-1883, October 2017

 

Please note : You will need Adobe Acrobat viewer to view the full articles.Get Free Adobe Reader

 
This site uses encryption for transmitting your passwords. ratmilwebsolutions.com