基于视觉感知的智能扫地机器人的垃圾检测与分类
Garbage detection and classification of intelligent sweeping robot based on visual perception
- 2019年24卷第8期 页码:1358-1368
收稿:2018-08-07,
修回:2019-3-17,
纸质出版:2019-08-16
DOI: 10.11834/jig.180475
移动端阅览

浏览全部资源
扫码关注微信
收稿:2018-08-07,
修回:2019-3-17,
纸质出版:2019-08-16
移动端阅览
目的
2
为了提高扫地机器人的自主性和智能化程度,为扫地机器人配备视觉传感器,使其获得视觉感知能力,通过研究有效的垃圾检测分类模型与算法,实现对垃圾的定位与识别,引导扫地机器人对垃圾进行自动识别与按类处理,提高工作的目的性和效率,避免盲动和减少能耗。
方法
2
选择检测速度较快的YOLOv2作为主网络模型,结合密集连接卷积网络,嵌入深层密集模块,对YOLOv2进行改进,提出一种YOLOv2-dense网络,该网络可以充分利用图像的高分辨率特征,实现图像浅层和深层特征的复用与融合。
结果
2
测试结果表明,智能扫地机器人使用本文方法可以有效识别不同形态的常见垃圾类别,在真实场景中,测试识别准确率为84.98%,目标检测速度达到26帧/s。
结论
2
实验结果表明,本文构建的YOLOv2-dense网络模型具有实时检测的速度,并且在处理具有不同背景、光照、视角与分辨率的图片时,表现出较强的适应和识别性能。在机器人移动过程中,可以保证以较高的准确率识别出垃圾的种类,整体性能优于原YOLOv2模型。
Objective
2
Home service robots have attracted widespread attention in recent years due to their close relationship with the daily lives of humans. Sweeping robots are the first home service robots that have entered the consumer market and are available extensively. At present
the intelligent sweeping robots on the market have only basic functions such as automatic path planning
automatic charging
and automatic obstacle avoidance
thereby greatly reducing the workload of housework
which is an important reason it is widely accepted by the market. Despite this situation
the current level of intelligence of sweeping robots remains low
and the main shortcomings are reflected in two aspects:First
high-level perception and discriminating ability toward the environment are lacking. For example
the behavior pattern adopted in the cleaning process is usually a random walk mode
and some higher intelligence sweeping robots may support simple path planning functions
such as taking a "Z" path. However
this function is generally classified under the "blind" mode
because the robot will perform cleaning activities whether or not garbage is present in its working path. Therefore
the work efficiency is low
and the energy consumption is greatly increased. Second
the current sweeping robots generally do not have the ability to distinguish the category of garbage. If garbage can be handled according to their correct category
then it will not only facilitate the sorting of garbage but also meet environmental protection requirements. Sweeping robots are equipped with visual sensors to achieve visual perception to improve their autonomy and intelligentization. A study of the effective classification model and algorithm of garbage detection can develop the process of location and garbage recognition
and a sweeping robot can be guided to automatically recognize and deal with different types of garbage
thereby improving the purpose and efficiency of its work. Moreover
the sweeping robot can avoid being in the blind state and reduce unnecessary energy consumption.
Method
2
The proposed garbage detection and classification method based on machine vision technology aims to improve the autonomous ability of the sweeping robot. This method selects the YOLOv2 network as the main network
which has a fast detection speed in the regression method. The YOLOv2 network is combined with the dense convolutional network (DenseNet) to make full use of the high-resolution features of detection objects. The shallow and deep features of object detection can be reused and fused by embedding deep dense modules to reduce the loss of feature information. Finally
a garbage detection and classification model is built by using a multiscale strategy to train the detection network. The self-built sample dataset is used in the training and testing process. To expand the sample size
the experiment uses the image data enhancement tool ImageDataGenerator provided by Keras to perform horizontal mirror flip
random rotation
cutting
zooming
and other types of processing. The training process of the YOLOv2-dense network is as follows:First
the original image is obtained through data enhancement
and the dataset obtained by the original image and the data enhancement is manually marked. Then
YOLOv2-dense training is performed with the dataset
and the training model is obtained. To adapt large dynamic changes in the scale of the object during the movement of the mobile robot
a multiscale training strategy is adopted for network training
and the detection model is obtained. The mobile robot experiment platform uses the omnidirectional motion of the Anycbot four-wheel-drive robot chassis
the AVR motion controller
and the NVIDIA Jetson TX2 of the 256-core GPU as the visual processor.
Result
2
The detection network model is trained and tested with our self-built dataset. Test results show that the improved network structure can retain more original picture information and enhance the extraction ability of target features
and this method can effectively identify the common types of garbage
including different forms of liquid and solid
and quickly and accurately mark out the location of the garbage. Real-time detection with different angles and distances presents good results. The accuracy of YOLOv2 detection is only 82.3%
and the speed is 27 frames per second. By contrast
the accuracy of the proposed improved network YOLOv2-dense reaches 84.98%
which is 2.62% higher than that of YOLOv2
and a speed of 26 frames per second can be achieved
which satisfies real-time detection.
Conclusion
2
The experimental results show that the YOLOv2-dense network model built in this study has the speed of real-time detection. It shows strong adaptability and recognition performance when dealing with pictures with different backgrounds
illuminations
visual angles
and resolutions. For the same object
the detection effect is greatly changed at different angles and scales
and the detection performance of long-distance objects is poor. Moreover
for 3D objects
different viewing angles have a greater impact on detection. The result of the object detection is dynamically changed during the movement of the robot. The proposed method can ensure that the type of garbage is accurately identified most of the time or within a certain distance and angle of view. Moreover
the proposed method can guarantee the identification of garbage types with higher accuracy
and its overall performance is better than that of the original YOLOv2 model.
Zhu D Q, Yan M Z. Survey on technology of mobile robot path planning[J]. Control and Decision, 2010, 25(7):961-967.
朱大奇, 颜明重.移动机器人路径规划技术综述[J].控制与决策, 2010, 25(7):961-967. [DOI:10.13195/j.cd.2010.07.4.zhudq.014]
Zhang H B, Yuan K, Zhou Q R. Visual navigation of a mobile robot based on path recognition[J]. Journal of Image and Graphics, 2004, 9(7):853-857.
张海波, 原魁, 周庆瑞.基于路径识别的移动机器人视觉导航[J].中国图象图形学报, 2004, 9(7):853-857. [DOI:10.11834/jig.200407161]
Yang J Y, Ma L, Bai D C, et al. Robot vision environmental perception method based on hybrid features[J]. Journal of Image and Graphics, 2012, 17(1):114-122.
杨俊友, 马乐, 白殿春, 等.机器人的混合特征视觉环境感知方法[J].中国图象图形学报, 2012, 17(1):114-122. [DOI:10.11834/jig.20120116]
Wang W G, Wu K B, Zhu J, et al. Obstacle detection for robot in unknown environment[J]. Journal of Image and Graphics, 2012, 17(2):209-214.
王文格, 武凯宾, 朱江, 等.未知环境下机器人障碍物检测技术[J].中国图象图形学报, 2012, 17(2):209-214. [DOI:10.11834/jig.20120208]
Gao Q J, Li J, Ma L, et al. Road crossing scene recognition for robot vision_based location[J]. Journal of Image and Graphics, 2009, 14(12):2510-2516.
高庆吉, 李娟, 马乐, 等.机器人视觉定位中的路口场景识别方法研究[J].中国图象图形学报, 2009, 14(12):2510-2516. [DOI:10.11834/jig.20091213]
Sun H, Wang C, Wang R S. A review of local invariant features[J]. Journal of Image and Graphics, 2011, 16(2):141-151.
孙浩, 王程, 王润生.局部不变特征综述[J].中国图象图形学报, 2011, 16(2):141-151. [DOI:10.11834/jig.20110207]
Lowe D G. Object recognition from local scale-invariant features[C]//Proceedings of the 7th IEEE International Conference on Computer Vision. Kerkyra, Greece: IEEE, 1999: 1150-1157.[ DOI: 10.1109/ICCV.1999.790410 http://dx.doi.org/10.1109/ICCV.1999.790410 ]
Bay H, Ess A, Tuytelaars T, et al. Speeded-up robust features (SURF)[J]. Computer Vision and Image Understanding, 2008, 110(3):346-359.[DOI:10.1016/j.cviu.2007.09.014]
Rublee E, Rabaud V, Konolige K, et al. ORB: an efficient alternative to SIFT or SURF[C]//Proceedings of 2011 International Conference on Computer Vision. Barcelona, Spain: IEEE, 2011: 2564-2571.[ DOI: 10.1109/ICCV.2011.6126544 http://dx.doi.org/10.1109/ICCV.2011.6126544 ]
Dalal N, Triggs B. Histograms of oriented gradients for human detection[C]//Proceedings of 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. San Diego, CA, USA: IEEE, 2005: 886-893.[ DOI: 10.1109/CVPR.2005.177 http://dx.doi.org/10.1109/CVPR.2005.177 ]
Krizhevsky A, Sutskever I, Hinton G E. ImageNet classification with deep convolutional neural networks[C]//Proceedings of the 25th International Conference on Neural Information Processing Systems. Lake Tahoe, Nevada: ACM 2012: 1097-1105.
Zheng Y, Chen Q Q, Zhang Y J. Deep learning and its new progress in object and behavior recognition[J]. Journal of Image and Graphics, 2014, 19(2):175-184.
郑胤, 陈权崎, 章毓晋.深度学习及其在目标和行为识别中的新进展[J].中国图象图形学报, 2014, 19(2):175-184. [DOI:10.11834/jig.20140202]
LeCun Y, Bengio Y, Hinton G. Deep learning[J]. Nature, 2015, 521(7553):436-444.[DOI:10.1038/nature14539]
Girshick R, Donahue J, Darrell T, et al. Rich feature hierarchies for accurate object detection and semantic segmentation[C]//Proceedings of 2014 IEEE Conference on Computer Vision and Pattern Recognition. Columbus, OH, USA: IEEE, 2014: 580-587.[ DOI: 10.1109/CVPR.2014.81 http://dx.doi.org/10.1109/CVPR.2014.81 ]
Girshick R. Fast R-CNN[C]//Proceedings of 2015 IEEE International Conference on Computer Vision. Santiago, Chile: IEEE, 2015: 1440-1448.[ DOI: 10.1109/ICCV.2015.169 http://dx.doi.org/10.1109/ICCV.2015.169 ]
Ren S Q, He K M, Girshick R, et al. Faster R-CNN:towards real-time object detection with region proposal networks[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2017, 39(6):1137-1149.[DOI:10.1109/TPAMI.2016.2577031]
Redmon J, Divvala S, Girshick R, et al. You only look once: unified, real-time object detection[C]//Proceedings of 2016 IEEE Conference on Computer Vision and Pattern Recognition. Las Vegas, NV, USA: IEEE, 2016: 779-788.[ DOI: 10.1109/CVPR.2016.91 http://dx.doi.org/10.1109/CVPR.2016.91 ]
Liu W, Anguelov D, Erhan D, et al. SSD: single shot MultiBox detector[C]//Proceedings of the 14th European Conference on Computer Vision. Amsterdam, The Netherlands: Springer, 2016: 21-37.[ DOI: 10.1007/978-3-319-46448-0_2 http://dx.doi.org/10.1007/978-3-319-46448-0_2 ]
Redmon J, Farhadi A. YOLO9000: better, faster, stronger[C]//Proceedings of 2017 IEEE Conference on Computer Vision and Pattern Recognition. Honolulu, HI, USA: IEEE, 2017: 6517-6525.[ DOI: 10.1109/CVPR.2017.690 http://dx.doi.org/10.1109/CVPR.2017.690 ]
Huang G, Liu Z, Van Der Maaten L, et al. Densely connected convolutional networks[C]//Proceedings of 2017 IEEE Conference on Computer Vision and Pattern Recognition. Honolulu, HI, USA: IEEE, 2017: 2261-2269.[ DOI: 10.1109/CVPR.2017.243 http://dx.doi.org/10.1109/CVPR.2017.243 ]
相关作者
相关机构
京公网安备11010802024621