| 注册
首页|期刊导航|农业工程学报|基于改进FasterR-CNN识别深度视频图像哺乳母猪姿态

基于改进FasterR-CNN识别深度视频图像哺乳母猪姿态

薛月菊 张南峰 朱勋沐 郑婵 毛亮 杨阿庆 涂淑琴 黄宁 杨晓帆 陈鹏飞

农业工程学报2018,Vol.34Issue(9):189-196,8.
农业工程学报2018,Vol.34Issue(9):189-196,8.DOI:10.11975/j.issn.1002-6819.2018.09.023

基于改进FasterR-CNN识别深度视频图像哺乳母猪姿态

Lactating sow postures recognition from depth image of videos based on improved Faster R-CNN

薛月菊 1张南峰 2朱勋沐 3郑婵 4毛亮 1杨阿庆 5涂淑琴 1黄宁 1杨晓帆 5陈鹏飞1

作者信息

  • 1. 华南农业大学电子工程学院,广州 510642
  • 2. 广东省现代养猪数据化工程技术研究中心,广州 510642
  • 3. 广东省农情信息监测工程技术研究中心,广州 510642
  • 4. 广州出入境检验检疫局,广州 510623
  • 5. 华南农业大学数学与信息学院,广州 510642
  • 折叠

摘要

Abstract

The maternal behaviors reflect the health and welfare of the sows, which directly affect the economic benefit of the pig farm. Computer vision provides an effective, low-cost and non-contact method for monitoring the behavior of animal for precision farming. Under the scene of piggery, it is a challenge for 24-hour automatic recognition of lactating sow postures due to the daily illumination variations, influence of heat lamp, and adhesion between piglets and sows. This paper proposed an automatic recognition algorithm of lactating sow postures based on improved Faster R-CNN (convolutional neural network) using depth video images. To improve the recognition accuracy and satisfy the real-time need, we designed a ZF-D2R (ZF with deeper layers and 2 residual learning frameworks) network by introducing residual learning frameworks into ZF network. First, 3 convolutional layers were added in the ZF network to design ZF-D (ZF with deeper layers). Then, in ZF-D network, shortcut connections were used to form 2 residual learning frameworks. The whole network made up the ZF-D2R network. Moreover, the Center Loss was introduced to Fast R-CNN detector to construct a joint classification loss function. With the joint supervision signals of F-SoftmaxLoss and Center Loss in Fast R-CNN detector, a robust model was trained to obtain the deep feature representations with the 2 key learning objectives, which led to intra-class compactness and inter-class dispersion as much as possible. So, the joint supervision of F-SoftmaxLoss and Center Loss could reduce recognition errors caused by the similar features between different postures. By taking ZF-D2R as basic net and adding the Center Loss to Fast R-CNN detector, the improved Faster R-CNN was built. Experiments to obtain the actual data set of lactating sow posture from the depth video of sows in the 28 pens were performed. The data set included 2 451 standing images, 2 461 sitting images, 2 488 sternal recumbency images, 2 519 ventral recumbency images and 2658 lateral recumbency images. And 5 000 images were randomly chosen as the testing set. The rest of the images were used as training set. To enhance the diversity of training data, dataset augmentation including rotating and mirroring was employed. Based on the Caffe deep learning framework, our improved Faster R-CNN was trained with end-to-end approximate joint methods. By adding 2 residual learning frameworks to ZF-D, the ZF-D2R model improved the MAP (mean of average precision) by 1.28 percentage points. After introducing the Center Loss supervision signal, the MAP of the optimal model reached 93.25%, obtaining an increase of 1.3 percentage points, and the MAP of the method proposed achieved 93.25%. And APs (average precisions) of the 5 classes of postures i.e. standing, sitting, sternal recumbency, ventral recumbency and lateral recumbency were 96.73%, 94.62%, 86.28%, 89.57% and 99.04%, respectively. The MAP of our approach was 3.86 and 1.24 percentage points higher than that of Faster R-CNN based on ZF basic net and Faster RCNN based on the deeper VGG16 basic net, respectively. Our method processed images at a speed of 0.058 s per frame, 0.034 s faster than Faster R-CNN based on VGG16. Our proposed method could improve the recognition accuracy and simultaneously ensure the real-time performance. Compared with DPM (deformable part model) detector plus CNN posture classifier, the MAP of the end-to-end recognition method proposed in this paper was increased by 37.87 percentage points, and the speed was raised by 0.855 s per frame. Our method can be used for the 24-hour recognition of sow behaviors and lays the foundation for the analysis of sow dynamic behavior by video.

关键词

图像识别/算法/模型/FasterR-CNN/残差结构/CenterLoss/哺乳母猪/姿态识别

Key words

image recognition/algorithms/models/Faster R-CNN/residual learning frameworks/Center Loss/lactating sow/posture recognition

分类

信息技术与安全科学

引用本文复制引用

薛月菊,张南峰,朱勋沐,郑婵,毛亮,杨阿庆,涂淑琴,黄宁,杨晓帆,陈鹏飞..基于改进FasterR-CNN识别深度视频图像哺乳母猪姿态[J].农业工程学报,2018,34(9):189-196,8.

基金项目

国家科技支撑计划(2015BAD06B03-3) (2015BAD06B03-3)

广东省科技计划项目(2015A020209148) (2015A020209148)

广东省应用型科技研发项目(2015B010135007) (2015B010135007)

广州市科技计划项目(201605030013) (201605030013)

广州市科技计划项目(201604016122) (201604016122)

农业工程学报

OA北大核心CSCDCSTPCD

1002-6819

访问量1
|
下载量0
段落导航相关论文