b83 Finalversion Esigelec WSCG 30-04-2019
b83 Finalversion Esigelec WSCG 30-04-2019
JY. Ertaud
Institute for Embedded Systems
Research, ESIGELEC,
UNIRouen, Normandy
University,
Saint Etienne du Rouvray,
76800, France
jean-yves.ertaud@esigelec.fr
ABSTRACT
This work aims to show the new approaches in our improved HOG approach and implemented two
embedded vision dedicated to object detection and kinds of PID controllers. The platform has been
tracking for drone visual control. Object/Pedestrian validated under different scenarios by comparing
detection has been carried out through two methods: measured data to ground truth data given by the drone
1. Classical image processing approach through GPS. Several tests which were ca 1rried out at
improved Histogram Oriented Gradient (HOG) and ESIGELEC car park and Rouen city center validate
Deformable Part Model (DPM) based detection and the developed platform.
pattern recognition methods. In this step, we present Keywords
our improved HOG/DPM approach allowing the Object detection, object recognition, visual tracking,
detection of a target object in real time. The developed tracking, pedestrian detection, deep learning, visual
approach allows us not only to detect the object servoing, HOG, DPM.
(pedestrian) but also to estimates the distance between
the target and the drone. 2. Object/Pedestrian
detection-based Deep Learning approach. The target 1. INTRODUCTION
position estimation has been carried out within image The works presented in this paper are a part of
analysis. After this, the system sends instruction to the ADAPT1 project (Assistive Devices for empowering
drone engine in order to correct its position and to disAbled People through robotic Technologies) which
track target. For this visual servoing, we have applied
Faster-DPM Algorithm
The DPM algorithm is applied on the entire image and
this is done at each iteration of the pedestrian video
sequence. In order to reduce the computing time, the
idea is to apply HOG algorithm only in a Region of
Interest (RoI) in which the target (here the pedestrian)
is located. This will drastically reduce the computing
time and will also better isolate the object. Firstly, the
DPM is applied all over the image once to locate the
object. Secondly, and after obtaining a RoI
Figure 5. Detected RoI surface (blue curve) vs system, the instructions are considered as speeds sent
distance from the object (red curve): y axis to drone engines.
represents the surface (pixels2), and x axis Firstly, we have applied a Proportional (P) controller:
represents distance (m).
Altitude : K p = 0.004
Forward translation : K p = 0.08
The blue curve looks like a square root. This is
Secondly, we have applied a classic PID controller (3):
validated by the trend curve, whose as equation (1) is:
t
Vrotation = K p ∗ erreur + K i ∗ ∫0 erreur ∗ dt +
−1.905
𝑦 = 262673 ∗ 𝑥 (1) Kd ∗
d
where x is the abscissa and y the ordinate. (erreur) (3)
dt
The equation of blue curve is (2):
9. REFERENCES
[1]. H. Cho, P. E. Rybski, A. B-Hillel, W. Zheng.:
Figure 8. Pedestrian Detection with improved Real-time Pedestrian Detection with Deformable
HOG/DPM approach under ESIGELEC Part Models, 2012
Indoor/Outdoor dataset. Left top image: [2]. Caltech Pedestrian Detection Benchmark
Pedestrian detection with calculated distance Homepage,
under DPM, Right top image: Pedestrian http://www.vision.caltech.edu/Image_Datasets/C
detection-based SSD deep learning model, Left altechPedestrians/, last accessed 2018/01/14.
bottom image: Pedestrian detection-based DPM, [3]. Felzenszwalb, P., McAllester, D., & Ramanan, D.
Right bottom image: Pedestrian detection-based A discriminatively trained, multiscale,
Faster-DPM. deformable part model. In Computer Vision and
Pattern Recognition, 2008. CVPR 2008. IEEE
Conference on (pp. 1-8). IEEE (June 2008).
7. CONCLUSION [4]. Jong-Chyi Su.: State of the Art object Detection
In this paper, we have presented a new approach for Algorithms. University of California, San Diego,
object detection and tracking applied for drone visual 9500 Gilman Dr. La Jolla, CA., 2014.
control. A comparison study between different [5]. Koen E. A. van de Sande, Jasper R. R. Uijlings,
approaches dedicated to pattern recognition and TheoGevers, Arnold W. M. Smeulders,
object/pedestrian detection has been carried out. We Segmentation As Selective Search for Object
have present our contribution to improve the quality Recognition, ICCV, 2011.
of both HOG and DPM algorithms. We have also [6]. Alex Krizhevsky , Ilya Sutskever , Geoffrey E.
implemented deep learning based Yolo and SSD Hinton. Imagenet classification with deep
algorithm for pedestrian detection. An innovative convolutional neural networks, NIPS, 2012.
method for distance calculation of pedestrian/object is [7]. Ross Girshick, Jeff Donahue, Trevor Darrell,
presented in this paper. The approach were validated Jitendra Malik, Rich feature hierarchies for
within experiments and the accuracy is 5% of errors. accurate object detection and semantic
The system detects not only object/target to follow but segmentation, CVPR, 2014.
also their distance from the drone. The system is [8]. P. Dollar, C. Wojek, B. Schiele.: Pedestrian
generic so that it is "applicable" on any type of Detection: An Evaluation of the State of the Art.
platform and/or environment (pedestrian detection for IEEE Transactions on Pattern Analysis and
autonomous vehicle and smart mobility, object Machine Intelligence,Volume: 34, Issue: 4, April
detection for smart wheelchair, object detection for 2012.
autonomous train, etc.). This work aims to show the [9]. D. G. Lowe. Distinct Image Features from Scale-
feasibility of scientific and technological concepts that Invariant Keypoints. Computer Science
affect object detection and tracking-based classic Department. University of British Columbia.
approaches like HOG/DPM, but also deep learning Vancouver, B. C. Canada. January, 5, 2004 28.
approaches-based Yolo or SSD. We have validated the [10]. Kanade-Lucas-Tomasi. KLT Feature
whole development under several scenarios by using Tracker. Computer Vision Lab. Jae Kyu Suhr.
both parrot drone platform and real vehicle in real Computer Vision (EEE6503) Fall 2009, Yonsei
traffic conditions (city center of Rouen in France). Uni.
[11]. J. M. Odobez and P. Bouthemy. Robust
Multiresolution Estimation of Parametric Motion
8. ACKNOWLEDGMENTS Models. IRIS/INRIA Rennes, Campus de
This research is supported by ADAPT project (This Beaulieu, February, 13, 1995.
work is carried out as part of the INTERREG VA [12]. Object Detection Homepage,
FMA ADAPT project ”Assistive Devices for http://cseweb.ucsd.edu/~jcsu/reports/ObjectDete
empowering disAbled People through robotic ction.pdf, last accessed 2018/01/14.
Technologies” http://adapt-project.com/index.php. [13]. Yan, J., Lei, Z., Wen, L., & Li, S. Z. The
The Interreg FCE Programme is a European fastest deformable part model for object
Territorial Cooperation programme that aims to fund detection. In Proceedings of the IEEE Conference
high quality cooperation projects in the Channel on Computer Vision and Pattern Recognition (pp.
border region between France and England. The 2497-2504). 2014.
[14]. Yolo Homepage, 2009 IEEE Conference on Computer Vision and
https://pjreddie.com/darknet/yolo/, last accessed Pattern Recognition. IEEE, Piscataway, NJ, pp.
2018/01/19. 304-311. ISBN 978-1-4244-3991-1. 2009.
[15]. Koen E. A. van de Sande, Jasper R. R. [26]. Junjie Yan, Xucong Zhang, Zhen Lei,
Uijlings, Theo Gevers, Arnold W. M. Smeulders, Shengcai Liao, Stan Z. Li. Robust Multi-
Segmentation As Se-lective Search for Object Resolution Pedestrian Detection in Traffic
Recognition, ICCV, 2011. Scenes. Computer Vision Foundation. CVPR
[16]. Alex Krizhevsky , Ilya Sutskever , Geoffrey 2013.
E. Hinton. Imagenet classification with deep [27]. http://www.bmw.com/com/en/insights/techn
convolutional neural networks. NIPS, 2012. ology/efficientdynamics/phase_2/, last accessed
[17]. Dave Gershgon, Google’s Robot Are 2018/01/14.
Learning How to Pick Things Up. Popular [28]. P. Viola, M. Jones. Rapid Object Detection
Science, March 8, 2016. using a Boosted Cascade of Simple Features.
[18]. Andreas Geiger, Philip Lenz, Christoph Computer Vision and Pattern Recongition
Stiller, Raquel Urtasun. Object Detection Conferencies. 2001.
Evaluation. The Kitti Vision Benchmark Suite. [29]. Mohamed, G., Ali, S. A., & Langlois, N.
Karlsruhe Institute of Technology. IEEE CVPR, (2017). Adaptive Super Twisting control design
2012. for manufactured diesel engine air path. The
[19]. Eddie Forson. Understanding SSD International Journal of Advanced Manufacturing
MultiBox-Real Time Object Detection in Deep Technology, 1-12. 2017.
Learning. Towards Data Science. November, [30]. ROS homepage:
18th. 2017. http://www.generationrobots.com/blog/fr/2016/0
[20]. Liu W., Anguelov D., Erhan D., Szegedy C., 3/ros-robot-operating-system-3/. Last accessed
Reed S., Fu C.-Y., Berg A. C., (2016, déc.). 2017/12/12.
Single-Shot Multibox Detector. [31]. B. Louvat, “Analyse d’image pour
https://arxiv.org/abs/1512.02325. l’asservissement d’une camera embarquée dans
[21]. Levine S., Pastor P., Krizhevsky A., Ibarz J., un drone”. Gipsa-lab, February, 5th, 2008,
Quillen D. (2017, June). Learning Hand-Eye Grenoble. 2008.
Coordination for Robotic Grasping with [32]. B. Hérissé, “Asservissement et navigation
DeepLearning and Large-Scale Data Collection. autonome d’un drone en environnement incertain
The International Journal of Robotics Research. par flot optique. PhD thesis, Université Sophia
[22]. Chabot F., Chaouch M., Rabarisoa J., Antipolis, UNSA et I2S CNRS, November 19th,
Teulière C., Château T.,. (2017, July). Deep 2010.
MANTA: A Coarse-to-fine Many-Task Network [33]. R. Girshick, “Fast R-CNN,” ICCV, 2015
for joint 2D and 3D vehicle analysis from [34]. S. Ren et al., “Faster R-CNN: Towards Real-
monocular image. IEEE Conference on Computer Time Object Detection with Region Proposal
Vision and Pattern Recognition. Networks,” arXiv:1506.01497. 2015.
[23]. Mousavian A., Anguelov D., Flynn J., [35]. H. Bay, T. Tuytelaars, and L. V. Gool.
Kosecka J. (2017, July). 3D Bounding Box “SURF: Speeded Up Robust Features”, Computer
Estimation Using Deep Learning and Geometry. Vision – ECCV 2006, pp. 404-417. 2006.
IEEE Conference on Computer Vision and [36]. D. Gerónimo, A. López, D. Ponsa, A.D.
Pattern Recognition. Sappa. “Haar wavelets and edge orientation
[24]. Georgakis G., Mousavian A., Berg A. C., histograms for on–board pedestrian detection”.
Kosecka J. (2017, July). Synthesizing Training In: Pattern Recognition and Image Analysis, pp.
Data for Object Detection in Indoor Scenes. IEEE 418-425. 2007.
Conference on Computer Vision and Pattern
Recognition
[25]. P. Dollar, C. Wojek, B. Schiele, and P.
Perona. Pedestrian Detection: A Benchmark. In: