A new multi-scale image semantic understanding method based on deep learning

Ying Feng Jiang, Hua Zhang*, Yan Bing Xue, Mian Zhou, Guang Ping Xu, Zan Gao

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

7 Citations (Scopus)

Abstract

How to fuse multi-scale information of image in deep learning is a key problem to be solved. To solve these problems, this paper proposed a deep learning method based on multi-scale iterative training for image semantic understanding. The algorithm uses the convolution neural network (CNN) to extract dense feature vectors from raw pixel for encoding regions centered on each pixel. The multi-scale iterative training captures different scales of textures, colors, edges and other important information. A new method combined with superpixel segmentation is proposed, to estimate the leading category of superpixel block and to correct the pixel classification error. It can depict the outline of the target area boundary and complete the final semantic understanding. The experiments on Stanford Background Dataset-8 verify the effectiveness of the proposed method, and the accuracy rate is 77.4%.

Original languageEnglish
Pages (from-to)224-230
Number of pages7
JournalGuangdianzi Jiguang/Journal of Optoelectronics Laser
Volume27
Issue number2
DOIs
Publication statusPublished - 15 Feb 2016
Externally publishedYes

Keywords

  • Convolutional neural network (CNN)
  • Deep learning
  • Image semantic understanding
  • Superpixel segmentation

Fingerprint

Dive into the research topics of 'A new multi-scale image semantic understanding method based on deep learning'. Together they form a unique fingerprint.

Cite this