Image Based Visual Servo Control UAV
Image Based Visual Servo Control UAV
Abstract— This paper proposes a control framework to it only allows an optimal trajectory in 3D Cartesian space
achieve the tracking of the moving target by a fixed-wing to be followed theoretically, instead of in the image space.
unmanned aerial vehicle (UAV) with a monocular pan-tilt As a result, even small errors in the image measurements
camera. This control framework is based on the image-based
visual servoing (IBVS) method, which takes the target feature can lead to errors in the pose [9], which may affect the
point on the captured image as the input and outputs the control tracking performance. Compared to PBVS, IBVS directly
signals directly with the aid of image Jacobian matrix. However, defines the difference between the current and desired image
the image is affected by the attitude of both the UAV and the characteristics as the control error, thereby eliminating the
pan-tilt, and the attitude of the pan-tilt is coupled with that process of 3D modeling, and insensitive to the calibration
of the UAV simultaneously. To solve this problem, we present
an “Ideal State” as the reference state, and make sure the errors of the sensors, cameras and robots. Therefore, this
coordinates of the feature point in the state are only affected kind of methods can better ensure that the target is in the
by the change of the yaw angle of the UAV. In this way, we can camera field of view [14].
integrate the attitude control of the UAV and the pan-tilt. By Recently, applying the IBVS technique to tackle the vision
using this control framework, the fixed-wing UAV can track the based control problem in UAVs is more and more common.
ground target continuously on the one hand, and the target will
tend to locate at image center on the other hand. This prevents Serra et al. [15], [16] use a quadrotor to track a moving plat-
the target from moving toward to the edge of the image or even form and then land on it with the aid of IBVS. Srivastava et
disappearing. Besides, we prove the controller is exponentially al. [17] control the quadrotor to keep following a target drone
convergent by the Lyapunov method. In order to evaluate the while maintaining a fixed distance from it in the absence of
performance of our controller, we build a hardware-in-the-loop GPS. Lyu et al. [18] propose a framework to realize vision-
(HIL) simulation platform and a prototype platform. Based
on these platforms, extensive experiments including simulations based multi-UAV cooperative mapping and control based on
and real flight tests are conducted. The results show that our IBVS. Although a number of approaches based on IBVS for
controller can achieve continuous and robust tracking of the UAVs emerge, most of the existing work is proposed for
target with a speed of 20km/h when the speed of the UAV is rotor UAVs [19], [20], [21], [22]. Compared to rotor UAVs,
16m/s. the fixed-wing UAVs have much more dynamical constraints,
I. INTRODUCTION such as it cannot move omni-direction on the one hand, and
its minimum speed is limited by the stalling speed on the
The past decade has witnessed a rapid development of other hand. Therefore, applying IBVS in target tracking for
UAVs. They can execute dull, dirty or dangerous tasks fixed-wing UAVs is more challenging.
instead of humans and have been widely used in both the To this end, Florent et al. [23] design a controller for the
civilian and military fields. Due to the excellent maneuver- fixed-wing UAV with a fixed camera to track a ground target.
ability of the UAVs, applying them in ground target tracking Their work enforces the trajectory of the UAV to converge to
has tremendous benefits [1], [2], [3], [4]. Considering that the a cone and a plane at the same time, and only the tracking of
fixed-wing UAVs have longer navigation time, larger payload stationary target is studied. Pietro et al. [24] achieve the target
and faster flight speed than rotor UAVs, the fixed-wing UAVs tracking for the fixed-wing UAV equipped with a pan-tilt
are more suitable for the long endurance tracking tasks. camera. They require the pan-tilt to tend to be perpendicular
Visual servo control is a commonly used tracking method to the body, which causes the yaw capability of the pan-
[5], [6], [7]. There are mainly two categories [8], [9], [10]: tilt cannot be fully utilized. As a result, the target may be
position-based visual servoing (PBVS) and image-based vi- out of sight when it moves fast. Wang et al. [25] present a
sual servoing (IBVS). PBVS requires the controller to be framework for tracking a mobile ground target using a fixed-
designed in 3D Cartesian space, thus the camera needs to wing UAV. Yet the design of guidance law in this approach
be calibrated for the transformation of coordinate frame. is based on target localization, which is easily affected by
Basically, most of the existing work of target tracking for the calibration of the camera. In our previous work [26], we
the UAVs adopt PBVS [11], [12], [13]. In this approaches, have studied the tracking of ground target by fixed-wing UAV
with fixed camera. Nevertheless, due to the limited sight of
This work was funded by the National Natural Science Foundation of
China (61906209) and (61973309) view and the fixed attitude of the camera, it is easy to lost
1 authors are with the College of Intelligence Science and the target when the target moves fast.
Technology, National University of Defense Technology, Changsha, In this work, we propose a control framework based on
China. ljyang13@163.com, zhliu@nudt.edu.cn,
guanzhengw@163.com, xkwang@nudt.edu.cn IBVS for tracking a moving target by a fixed-wing UAV
* corresponding author with a pan-tilt camera. This approach integrates the attitude
355
Authorized licensed use limited to: DELHI TECHNICAL UNIV. Downloaded on March 12,2025 at 18:45:41 UTC from IEEE Xplore. Restrictions apply.
Fig. 3: The range of the pan-tilt attitude. Fig. 4: The framework of the controller.
where uψ , u p and ut are control inputs. The range of the detection is not the focus of the paper, we adopt existing
pan-tilt attitude is shown as Fig. 3, γ p1 (< 0), γ p2 (> 0) and method named YOLOv3 [29] to solve the problem directly.
γt1 (< 0) are the thresholds of the pan-tilt attitude. 2) Servo Control: Image-based visual servoing is able to
Different from the conventional position-based tracking directly map the desired velocity of feature point into the
method, we intend to propose an IBVS control law for the motion velocity of the UAV. With the aid of it, there is no
UAV and the pan-tilt together by using the image information need to calculate the relative position of the target to the
directly, without target positioning processes in FC . However, UAV. As a result, the generated error in this process can be
there exist two challenges in the work as follows. eliminated.
We denote the desired velocity of the centroid coordinates
• A small jitter of the UAV attitude will cause a dramatical
by Ṡ(u̇, v̇) in FC , and it is related with the altitude of the
jump for the position of the target in the image, and
UAV and the pan-tilt. Simultaneously, the linear velocity of
the attitude of the fixed-wing UAV is dynamically
the UAV is denoted by T = (Vx ,Vy ,Vz )T , and angular velocity
changed all the time during the mission. Therefore, how
of it is denoted by Ω = (ωx , ωy , ωz )T . There exists image
to implement the tracking under this circumstance is
Jacobian matrix Jv that associates Ṡ with {T, Ω}, which is
challenging.
shown as Eq. (2).
• Both the attitude of the pan-tilt and that of the UAV are Å ã Å ã
coupled, how to integrate the attitude control of them is u̇ T
= Jv · , (2)
challenging as well. v̇ Ω
where Jv ∈ R2×6 [9], and we can obtain the control output
B. Overall Framework
directly from Ṡ through the inverse of the equation. Simul-
In this subsection, we design the overall framework of taneously, we need to adjust θ p and θt to make (u1 , v1 ) tend
the visual control scheme (Fig .4). The pitch, roll and yaw to (0, 0).
angle of the UAV in current state are denoted by {θ , ϕ , ψc }
in order, and the desired yaw angle of the UAV is denoted III. MAIN WORK
by ψe . Besides, H represents the flight height, α represents A. Control Law Design
a desired pitch angle of the pan-tilt, and kψ is a coefficient. During the target tracking process, the attitude of the UAV
Firstly, an image detection module is used to obtain the {ψ , θ , ϕ } is coupled with that of the pan-tilt {θ p , θt }. As a
target information, which is the target position (u1 , v1 ) in FI . result, the centroid coordinates of the target are affected by
And then, the servo control module is performed with the the attitude of both the UAV and the pan-tilt. Furthermore,
aid of the (u1 , v1 ) information and the UAV’s state, to obtain the coordinates of the desired feature point changes at
the desired uψ , u p and ut . runtime before the UAV reaches a stable loitering state,
1) Image Detection: UAV-based image detection is in which causes the unknown of Ṡ.
charge of processing image and detecting target. It takes In order to make Ṡ unaffected by the change of
the image captured by the airborne camera as the input, by {θ , ϕ , θ p , θt }, we propose a reference state named “Ideal
locating the position of the target on the image, and outputs State” (Definition 1). Through this method, we map the
the coordinates of the target feature point (denoted by (u1 , v1 ) current system state into an reference state, and Ṡ in the latter
in FI ). It should be noted that high requirement of accuracy can be determined uniquely. Moreover, Ṡ is only affected by
and real-time performance is necessary in field experiments. the change of ψ , then uψ can be obtained on the basis.
To the best of our knowledge, there exist two kinds of image Definition 1 (Ideal State): The “UAV & pan-tilt” system
detection methods [28], which can be divided into traditional that the model corresponding to must simultaneously satisfy
method and neural network method. Compared to the former, the following four conditions: (a) The pitch angle of the UAV
the latter can obtain (u1 , v1 ) directly from the image, which is is 0◦ ; (b) The roll angle of the UAV is 0◦ ; (c) The pitch angle
more efficient. Besides, considering that this part of image of the pan-tilt is α ; (d) The yaw angle of the pan-tilt is 90◦ .
356
Authorized licensed use limited to: DELHI TECHNICAL UNIV. Downloaded on March 12,2025 at 18:45:41 UTC from IEEE Xplore. Restrictions apply.
Fig. 5: Flow chart of the transformation for the camera
coordinate frame.
357
Authorized licensed use limited to: DELHI TECHNICAL UNIV. Downloaded on March 12,2025 at 18:45:41 UTC from IEEE Xplore. Restrictions apply.
Algorithm 1 Using coordinates of the feature point to obtain
the control of the UAV and the pan-tilt
Require: the image captured by camera
Ensure: the control of UAV and pan-tilt
1: while discover the target do
2: detect the centroid coordinates (u1 , v1 );
3: calculate the rotation matrices R1 , R2 , R3 , R4 , R5 ;
4: obtain the transformed centroid coordinates:
Ñ é
Å ã u1
u2
= k · (R5 · R4 · R3 · R2 · R1 )(1−2,·) v1
v2 Fig. 7: The HIL simulation environment.
−f
5: obtain the velocity of the transformed target:
Å ã Å ãÅ ã Set a Lyapunov function
u˙2 λ1 u2
=−
v˙2 λ2 v2 1
V = eT · e,
6: calculate M1 and M2 : 2
(1,5) (1,6)
it is obvious that V > 0, and the time derivative of V can be
M1 = −Jv · sinα + Jv · cosα represented as
(2,5) (2,6)
M2 = −Jv · sinα + Jv · cosα V̇ = eT · e = eT · Pe · e, (10)
7: obtain the yaw rate of the UAV: Since Pe is negative, and we choose
358
Authorized licensed use limited to: DELHI TECHNICAL UNIV. Downloaded on March 12,2025 at 18:45:41 UTC from IEEE Xplore. Restrictions apply.
120 50
u (pixel)
110
100 0
90
Distance (m)
80 -50
0 20 40 60 80 100
70
Time (s)
60
50
50
v (pixel)
40
0
30
20
-50
10 0 20 40 60 80 100
0 20 40 60 80 100 120
Fig. 8: The fixed-wing UAV tracks the target of uniform linear motion in the HIL simulation. The speed of the target is
4m/s, the flight height and speed of the UAV are 100m and 16m/s, respectively. (a) Tracking trajectories; (b) Horizontal
distance between the target and the fixed-wing UAV; (c) Image stabilizing precision.
359
Authorized licensed use limited to: DELHI TECHNICAL UNIV. Downloaded on March 12,2025 at 18:45:41 UTC from IEEE Xplore. Restrictions apply.
stationary, the angles of the pan-tilt cannot keep constant for
the unstable state of the fixed-wing UAV in field experiments.
Besides, the points inside blue circles deviate significantly,
which is caused by misdetection. And there are also some
points missing inside blue rectangles. The reasons accounting
for this can be divided into two kinds, one is that the
target is too small in the image edge to be detected, the
other is the angle of the pan-tilt reaches threshold, which
Fig. 10: The fixed-wing UAV and the car (as the ground makes the target out of sight briefly. Under the circumstance,
target) we use in the field experiments. we use the uψ calculated from (u2 , v2 ) at the last moment
before the target is lost for control. After that, the UAV will
yaw to the direction where the target disappears from the
image. As a result, the UAV can still track the target despite
these situations, which further verifies the feasibility of our
controller in real flight tests.
V. CONCLUSION
This paper has presented a control framework for a fixed-
wing UAV with a monocular pan-tilt camera to track a
moving target. More specifically, this control framework
uses the target detection algorithm based on YOLOv3 to
obtain the centroid coordinates of the target firstly. Then,
this framework uses a reference state called “Ideal State” to
make the coordinates of the feature point only affected by the
change of the yaw angle of the UAV. After that, the controller
is designed based on the image Jacobian matrix. In order to
Fig. 11: The scenes of the real flight tests. The red circle verify the feasibility of the controller, we have conducted the
represents the flying fixed-wing UAV, and the blue box shows HIL simulation experiments based on Gazebo, and then have
the target, which is a white car. carried out field experiments based on a prototype system.
The results show that our controller can achieve continuous
and robust tracking of the target by the UAV.
further evaluate our controller, and it is developed based on Future work will focus on target tracking by multiple
the existing architecture of our team [30], [31]. fixed-wing UAVs. We believe that by using the images
of diverse perspectives and surrounded trajectories from
The fixed-wing UAV and the car we use in the field
different UAVs cooperatively, the challenge of tracking the
experiments are shown in Fig. 10. The UAV has a wingspan
high speed maneuvering ground target can be met.
of 1.8m and a weight of 5kg. Besides, the length and height
of the UAV are 1.23m and 0.35m, respectively. For the pan- R EFERENCES
tilt camera, it is the same as we use in the HIL simulation
platform. [1] Wu J, Wang H, Li N, et al. Distributed trajectory optimization for
multiple solar-powered UAVs target tracking in urban environment by
Fig. 11 displays the scenes of our field experiments. As Adaptive Grasshopper Optimization Algorithm[J]. Aerospace Science
the figure shows, our experimental site is a meadow, and the and Technology, 2017, 70: 497-510.
target we use is a white car (the blue box). The four pictures [2] Zhang M, Liu H H T. Cooperative tracking a moving target using mul-
tiple fixed-wing UAVs[J]. Journal of Intelligent & Robotic Systems,
represent a sequence of scenes from the same experiment. 2016, 81(3-4): 505-529.
2) Experimental Results and Analysis: We make the [3] Vanegas F, Campbell D, Roy N, et al. UAV tracking and following
fixed-wing UAV track target with uniform linear motion in a ground target under motion and localisation uncertainty[C]//2017
IEEE Aerospace Conference. IEEE, 2017: 1-10.
the experiment. The flight airspeed and height are 16m/s [4] Vanegas F, Gonzalez F. Uncertainty based online planning for UAV
and 100m, respectively. Besides, the target is moving at a target finding in cluttered and GPS-denied environments[C]//2016
constant speed of 20km/h. IEEE Aerospace Conference. IEEE, 2016: 1-9.
[5] Larouche B P, Zhu Z H. Autonomous robotic capture of non-
The tracking trajectories are shown as Fig. 12a, and the cooperative target using visual servoing and motion predictive con-
horizontal distance between the UAV and the target is shown trol[J]. Autonomous Robots, 2014, 37(2): 157-167.
as Fig. 12b. We can find that the horizontal distance keeps [6] Pestana J, Sanchez-Lopez J L, Campoy P, et al. Vision based gps-
denied object tracking and following for unmanned aerial vehi-
at about 95m when the target is stationary, and it will vary cles[C]//2013 IEEE International Symposium on Safety, Security, and
within the range of [23, 210] (denoted by two black dashed rescue Robotics (SSRR). IEEE, 2013: 1-6.
lines). [7] Olivares Mendez M A, Mondragon I, Campoy P, et al. Aerial object
following using visual fuzzy servoing[C]//First Workshop on Research,
Fig. 12c shows the image stabilizing precision for the Development and Education on Unmanned Aerial Systems (RED-UAS
experiment. The green part indicates that even the target is 2011). 2011.
360
Authorized licensed use limited to: DELHI TECHNICAL UNIV. Downloaded on March 12,2025 at 18:45:41 UTC from IEEE Xplore. Restrictions apply.
250
step 1
step 2
200
Distance (m)
150
100
50
0
0 20 40 60 80 100 120
Time (s)
(a) (b) (c)
Fig. 12: The fixed-wing UAV tracks the target of uniform linear motion in the field experiment. The speed of the target is
20km/h, the flight height and speed of the UAV are 100m and 16m/s, respectively. Besides, the target is stationary in step
1 and moving in step 2. (a) Tracking trajectories; (b) Horizontal distance between the target and the fixed-wing UAV; (c)
Image stabilizing precision.
[8] Hutchinson S, Hager G D, Corke P I. A tutorial on visual servo inspection of a ground target using a multirotor UAV system[J].
control[J]. IEEE Transactions on Robotics and Automation, 1996, Sensors, 2017, 17(12): 2929.
12(5): 651-670. [22] Wei Y, Lin Z. Vision-based Tracking by a Quadrotor on ROS[J]. IFAC-
[9] Chaumette F, Hutchinson S. Visual servo control. I. Basic ap- PapersOnLine, 2017, 50(1): 11447-11452.
proaches[J]. IEEE Robotics & Automation Magazine, 2006, 13(4): [23] Le Bras F, Hamel T, Mahony R. Image-based visual servo control for
82-90. circular trajectories for a fixed-wing aircraft[C]//Proceedings of the
[10] Chaumette F, Hutchinson S. Visual servo control. II. Advanced ap- 48h IEEE Conference on Decision and Control (CDC) held jointly
proaches [Tutorial][J]. IEEE Robotics & Automation Magazine, 2007, with 2009 28th Chinese Control Conference. IEEE, 2009: 3430-3435.
14(1): 109-118. [24] Peliti P, Rosa L, Oriolo G, et al. Vision-based loitering over a target
[11] Teulire C, Eck L, Marchand E, et al. 3D model-based tracking for for a fixed-wing UAV[J]. IFAC Proceedings Volumes, 2012, 45(22):
UAV position control[C]//2010 IEEE/RSJ International Conference on 51-57.
Intelligent Robots and Systems. IEEE, 2010: 1084-1089. [25] Wang X, Zhu H, Zhang D, et al. Vision-based detection and tracking
of a mobile ground target using a fixed-wing UAV[J]. International
[12] Wise R, Rysdyk R. UAV coordination for autonomous target track-
Journal of Advanced Robotic Systems, 2014, 11(9): 156.
ing[C]//AIAA Guidance, Navigation, and Control Conference and
[26] Yang L, Liu Z, Wang X, et al. An optimized image-based visual servo
Exhibit. 2006: 6453.
control for fixed-wing unmanned aerial vehicle target tracking with
[13] Chen J, Dawson D M. UAV tracking with a monocular cam- fixed camera[J]. IEEE Access, 2019, 7: 68455-68468.
era[C]//Proceedings of the 45th IEEE Conference on Decision and [27] Zhao S, Wang X, Lin Z, et al. Integrating Vector Field Approach and
Control. IEEE, 2006: 3873-3878. Input-to-State Stability Curved Path Following for Unmanned Aerial
[14] Yongchun F. A survey of robot visual servoing [J]. CAAI Transactions Vehicles[J]. IEEE Transactions on Systems, Man, and Cybernetics:
on Intelligent Systems, 2008, 3(2): 109-114. Systems, 2018:1-8.
[15] Serra P , Cunha R , Hamel T , et al. Landing of a Quadrotor on a [28] Zhiqiang W, Jun L. A review of object detection based on convo-
Moving Target Using Dynamic Image-Based Visual Servo Control[J]. lutional neural network[C]//2017 36th Chinese Control Conference
IEEE Transactions on Robotics, 2017, 32(6):1524-1535. (CCC). IEEE, 2017: 11104-11109.
[16] Araar O , Aouf N , Vitanov I . Vision Based Autonomous Landing [29] Redmon J, Farhadi A. Yolov3: An incremental improvement[J]. arXiv
of Multirotor UAV on Moving Platform[J]. Journal of Intelligent & preprint arXiv:1804.02767, 2018.
Robotic Systems, 2017, 85(2):369-384. [30] Wang X, Shen L, Liu Z, et al. Coordinated flight control of miniature
[17] Srivastava R, Lima R, Das K, et al. Least Square Policy Iteration for fixed-wing UAV swarms: methods and experiments[J]. Science China
IBVS based Dynamic Target Tracking[C]//2019 International Confer- Information Sciences, 2019, 62(11): 212204.
ence on Unmanned Aircraft Systems (ICUAS). IEEE, 2019: 1089- [31] Liu Z, Wang X, Shen L, et al. Mission Oriented Miniature Fixed-wing
1098. UAV Swarms: A Multi-layered and Distributed Architecture[J]. arXiv
[18] Lyu Y, Pan Q, Zhang Y, et al. Simultaneously multi-UAV mapping preprint arXiv:1912.06285, 2019.
and control with visual servoing[C]//2015 International Conference
on Unmanned Aircraft Systems (ICUAS). IEEE, 2015: 125-131.
[19] Popova M G. Visual Servoing for a Quadrotor UAV in Target Tracking
Applications[M]. University of Toronto, 2015.
[20] Zarudzki M, Shin H S, Lee C H. An image based visual servoing ap-
proach for multi-target tracking using an quad-tilt rotor UAV[C]//2017
International Conference on Unmanned Aircraft Systems (ICUAS).
IEEE, 2017: 781-790.
[21] Hinas A, Roberts J M, Gonzalez F. Vision-based target finding and
361
Authorized licensed use limited to: DELHI TECHNICAL UNIV. Downloaded on March 12,2025 at 18:45:41 UTC from IEEE Xplore. Restrictions apply.