
2 International Journal of Automation and Computing
vehicles with regard to the expected time span for future
prediction. We classify the prediction of intention as long-
term intention prediction and short-term prediction. Fig. 1
demonstrates the surveying flowchart in this paper. At dis-
tinct stages, we discuss open problems and challenges, and
endeavour to provide possible solutions.
Fig. 1 The scene understanding flowchart by event reasoning
framework for autonomous driving
Actually, beyond those stages, some end-to-end ap-
proaches emerge recently for scene understanding facing
autonomous driving
[12−14]
. They rely on a large-scale data-
driven mechanism, and formulate the scene to decide with
deep layers or recursive perception, such as fast recurrent
fully convolutional networks (FCN) for direct perception in
autonomous driving
[12]
and FCN-LSTM
[13]
for a future mo-
tion action feasibility distribution. We specially take a sec-
tion to present this category. We hope that our survey can
sweep away some entry barriers of deep scene understanding
for autonomous driving, and draw forth meaningful insights
and solutions for this field.
1.1 Autonomy pursuit in driving
Developing autonomous systems aim to assist humans
in handing everyday tasks. Autonomous driving system, a
system for closely related to humans
everyday trips, has
become people
s one of the most typical pursuits. It can
free hands from the steering wheel, and spare time for tack-
ling many other things. Meanwhile, the equipped sensors of
autonomous vehicle can also recognize the surrounding con-
dition immediately and ensure safe driving, thus decreasing
traffic accidents. Encouraged by those merits, researchers
are diligently pursuing autonomous driving all the time.
There are two kinds of driving force in the development
of autonomous driving. One is the projects launched and
challenges posed by different governments, research insti-
tutes and vehicle manufacturers. The other we want to
emphasize is the publicly available benchmarks.
Projects and launched challenges. Since 1986, Eu-
rope started an intelligent transportation system project,
named as PROMETHEUS, involving more than 13 ve-
hicle manufacturers and research institutions around 19
countries. Thorpe et al.
[15]
in Carnegie Mellon University
launched the first autonomous driving project in the United
States. This project made breakthrough in 1995 that au-
tonomously drove a car from Pittsburgh, Pennsylvania to
San Diego, California. Supported by many related stud-
ies, the US government established the National Automated
Highway System Consortium (NAHSC) in 1995. Motivated
by these projects, highway scenarios has been intensively
studied for a long time, while urban scene remained as an
uncultivated area. Actually, urban scene is closely related
to human
s daily lives. At that time, a famous “DARPA
Grand Challenge (DUC)” launched by Defense Advanced
Research Projects Agency (DARPA) largely accelerated
the progress of autonomous vehicle. Among them, “Ur-
ban Challenge”
[16]
, the third challenge launched by DARPA
(others had been held in 2004 and 2005 respectively, aiming
to test the self-driving performance in the Mojave Desert of
the United States
[17, 18]
, took place on November 3, 2007 at
the now-closed George Air Force Base in Victorville, Cali-
fornia. Rules included obeying all traffic regulations while
negotiating with other vehicles and obstacles and merg-
ing into traffic. There were 4 teams completed the route
within 6 hours. In 2009, National Natural Science Foun-
dation of China launched the China Intelligent Vehicle Fu-
ture Challenge (iVFC). Up to now, the ninth contest was
held in November 2017. Google started their self-driving
car project in 2009, and completed over 5 million miles
driving test until March 2018
1
. In 2016, the project was
evolved into an independent self-driving technology com-
pany Waymo. Tesla Autopilot
2
, by equipping cameras,
twelve ultrasonic sensors and a forward-facing radar, all the
vehicles can have the self-driving ability since October 2016.
As a matter of fact, more and more vehicle manufacturers,
such as Audi, BMW, Benz, have begin their projects to
develop their self-driving vehicles.
Benchmarks. In 2012, Geiger et al.
[19]
introduced the
KITTI vision benchmark, which contained six different ur-
ban scenes, and had 156 video sequences with time span
from 2 minutes to 8 minutes. Within this benchmark,
they launched several typical vision tasks, such as pedes-
trian/vehicle detection, optical flow, stereo flow, road detec-
tion, lane detection, etc. The benchmark was collected by
an ego-vehicle equipped with color and gray cameras, and
Velodyne 3D laser scanner and high-precision GPS/IMU
inertial navigation systems. At the same time, Cambridge
University released CamVid dataset
[20]
, which provided a
semantic segmentation evaluation benchmark containing
only four video sequences on urban scene. Another pop-
ular benchmark is the Cityscapes dataset
[21]
released in
2016. Urban scene was collected in 50 cities, and have 5 000
fine-annotated images and 20 000 coarse-annotated images.
Cityscapes has become the most challenging dataset for
semantic segmentation task. Actually, annotation is time
and labor consuming. Based on that, Gaidon et al.
[22]
con-
structed a large-scale KITTI-like virtual dataset
3
by com-
puter graphic technology. The benefit of virtual dataset
is that it can generate every wanted task, for those that
1
https://www.theverge.com/2018/2/28/17058030/waymo-self-
driving-car-360-degree-video
2
https://www.tesla.com/autopilot
3
http://www.europe.nav erlabs.com/Research/Computer-
Vision/Proxy-Virtual-Worlds
评论