暂无图片
暂无图片
暂无图片
暂无图片
暂无图片
DeepSignals- Predicting Intent of Drivers Through Visual Signals.pdf
186
8页
0次
2021-05-01
50墨值下载
DeepSignals: Predicting Intent of Drivers Through Visual Signals
Davi Frossard
1,2
Eric Kee
1
Raquel Urtasun
1,2
Abstract Detecting the intention of drivers is an essential
task in self-driving, necessary to anticipate sudden events like
lane changes and stops. Turn signals and emergency flashers
communicate such intentions, providing seconds of potentially
critical reaction time. In this paper, we propose to detect these
signals in video sequences by using a deep neural network
that reasons about both spatial and temporal information. Our
experiments on more than a million frames show high per-frame
accuracy in very challenging scenarios.
I. INTRODUCTION
Autonomous driving has risen as one of the most impactful
applications of Artificial Intelligence (AI), where it has the
potential to change the way we live. Before self-driving cars
are the norm however, humans and robots will have to share
the roads. In this shared scenario, communications between
vehicles are critical to alert others of maneuvers that would
otherwise be sudden or dangerous. A social understanding
of human intent is therefore essential to the progress of self-
driving. This poses additional complexity for self-driving
systems, as such interactions are generally difficult to learn.
Drivers communicate their intent to make unexpected
maneuvers in order to give warning much further in advance
than would otherwise be possible to infer from motion. Al-
though driver movements communicate intent—for example
when drivers slow down to indicate that they will allow a
merge, or drive close to a lane boundary to indicate a desired
merge position—motion cues are subtle, context dependent,
and near-term. In contrast, visual signals, and in particular
signal lights, are unambiguous and can be given far in
advance to warn of unexpected maneuvers.
For example, without detecting a turn signal, a parked car
may appear equally likely to remain parked as it is to pull
into oncoming traffic. Analogously, when a driver plans to
cut in front of another vehicle, they will generally signal
in advance for safety. Buses also signal with flashers when
making a stop to pick up and drop off passengers, allowing
vehicles approaching from behind to change lanes, therefore
reducing delays and congestion.
These everyday behaviors are safe when drivers under-
stand the intentions of their peers, but are dangerous if visual
signals are ignored. Humans expect self-driving vehicles to
respond. We therefore consider in this work the problem
of predicting driver intent through visual signals, and focus
specifically on interpreting signal lights.
Estimating the state of turn signals is a difficult problem:
The visual evidence is small (typically only a few pixels),
1
Uber Advanced Technologies Group
2
University of Toronto
{frossard, ekee, urtasun}@uber.com
Fig. 1: A vehicle, signaling left, passes through occlusion.
The actor’s intent to turn left is correctly detected (left
arrow), including the occlusion (question mark).
particularly at range, and occlusions are frequent. In addition,
intra-class variations can be large. While some regulation
exists, many vehicles have stylized blinkers, such as light
bars with sequential lights in the direction being signaled,
and the regulated frequency of blinking (1.5 ± 0.5 Hz [1])
is not always followed. Furthermore, since we are interested
in estimating intent, vehicle pose needs to be decoded. For
instance, a left turn signal would correspond to a flashing
light on the left side of a vehicle we are following, but on
the other hand would correspond to a flashing light on the
right side of an incoming vehicle. We refer the reader to
Figure 2 for an illustration of some of the challenges of turn
signal estimation.
Surprisingly little work in the literature has considered
this problem. Earlier published works [2], [3] use hand-
engineered features, trained in-part on synthetic data, and
are evaluated on limited datasets. Other approaches have
considered only nighttime scenarios [4], [5]. Such methods
are unlikely to generalize to the diversity of driving scenarios
that are encountered every day.
In this paper, we identify visual signal detection as an
important problem in self-driving. We introduce a large-
scale dataset of vehicle signals, and propose a modern deep
learning approach to directly estimate turn signal states from
diverse, real-world video sequences. A principled network is
designed to model the subproblems of turn signal detection:
attention, scene understanding, and temporal signal detection.
This results in a differentiable system that can be trained end-
to-end using deep learning techniques, rather than relying
upon hard coded premises of how turn signals should behave.
We demonstrate the effectiveness of our approach on a
new, challenging real-world dataset comprising 34 hours of
加入知识星球 行业与管理资源库,免费下载报告合集
1.
上传分享2000+份最新行业资源(涵盖科、金融、教育、互联网、
房地、生物制药、医疗健康等行研报告、科技动态、管理方案等);
2. 免费下载资源库已存行业报告
3. 免费下载资源库已存国内外咨询公司管理方案企业运营制度
4. 免费下载资源库已存科技方案、论、报告及课件
微信扫码二维码,免费报告轻松领
微信扫码加入知识星球 行业与管理资源
获取更多行业报告、管理文案、大师笔记
加入微信群,每日获取免费3+份报告
1.
扫一扫二维码,添加客服微信(微信号Teamkon2
2.
添加好友请备注:姓名+单位+业务领域
3. 群主将邀请您进专业行业报告资源群
of 8
50墨值下载
【版权声明】本文为墨天轮用户原创内容,转载时必须标注文档的来源(墨天轮),文档链接,文档作者等基本信息,否则作者和墨天轮有权追究责任。如果您发现墨天轮中有涉嫌抄袭或者侵权的内容,欢迎发送邮件至:contact@modb.pro进行举报,并提供相关证据,一经查实,墨天轮将立刻删除相关内容。

评论

关注
最新上传
暂无内容,敬请期待...
下载排行榜
Top250 周榜 月榜