[关闭]
@ArrowLLL 2018-02-28T21:55:03.000000Z 字数 5760 阅读 2770

Study-Note: Soft + Hardwired Attention : An LSTM Framework for Human Trajectory Prediction and Abnormal Event Detection

Study-Note OPTIMAL


中文理解

首先,作者利用轨迹的起点和终点(source and sink position)将数据集场景中的轨迹进行聚类,

image_1c7ea23rg2it7cdgrg18i0143q4k.png-640.1kB

将人群分类后,对每一个类别训练不同的参数创建对应的轨迹模型。

image_1c7dn1b91151jc89mid0ook2o.png-309kB

作者的模型可以用上图来描述。首先使用LSTM encoder对当前时刻 之前 的轨迹点信息进行encoder,并获得相应的状态信息,然后使用注意力机制将每一个轨迹的信息再做一次聚合,对当前要预测的这个人使用软性注意力聚合(soft attention),即其系数由feed forward network 结合预测系统得到;而对于其周围的人的系数则根据距离这一硬性注意力(hardwired attention)聚合()得到。再通过一个全连接层将两种聚合产物()聚合为作为LSTM Decoder的输入,通过LSTM Decoder解析出预测的路径。

Abstract

we propose a novel method to predict the future motion of a pedestrian given a short history of their, and their neighbours, past behaviour. The novelty of the proposed method is the combined attention model which utilises both “soft attention” as well as “hard-wired” attention in order to map the trajectory information from the local neighbourhood to the future positions of the pedestrian of interest.

Introduction

The approach we present in this paper can be viewed as a data driven approach which learns the relationship between neighbouring trajectories in an unsupervise manner.

Problem Definition

The task we are interested in is predicting the trajectory of the pedestrian for the period of to , having observed the trajectory of the pedestrian from time to as well as the trajectories all the other pedestrians in the local neighbourhood during that period.

Proposed Solution

image_1c7dn1b91151jc89mid0ook2o.png-309kB

A sample surveillance scene (on the left): The trajectory of the pedestrian of interest is shown in green, and has two neighbours (shown in purple) to the left, one in front and none on right.

Neighbourhood encoding scheme (on the right): Trajectory information is encoded with LSTM encoders. A soft attention context vector is used to embed the trajectory information from the pedestrian of interest, and a hardwired attention context vector is used for neighbouring trajectories. In order to generate we use a soft attention function denoted in the above figure, and the hardwired weights are denoted by . The merged
context vector is then used to predict the future trajectory for the pedestrian of interest (shown in red).

Proposed Approach

LSTM encoder-decoder Framework

shAttention.PNG-122.4kB

The proposed Soft+Hardwired Attention model

We utilise the trajectory information from both the pedestrian of interest and the neighbouring trajectories. We embed the trajectory information from the pedestrian of interest with the soft attention context vector , while neighbouring trajectories are embedded with the aid of a hardwired attention context vector . In order to generate we use a soft attention function denoted in the above figure, and the hardwired weights are denoted by . Then the merged context vector, , is used to predict the future state

LSTM Encoder

The input sequence for the pedestrian is given by

The encoded sequence is given by,

The encoding function is an LSTM, which can be denoted by,

With the aid of above equation we encode tge trajectory information from hte pedestrian of interest as well as each trajectory in the local neighbourhood.

LSTM Decoder

The soft attention context vector can be computed as a weighted sum of hidden states,

The weight can be computed by,


the function is a feed forward network for joint training with other components of the system.

Let there be neighbouring trajectories in the local neighbourhood and be
the encoded hidden state of the neighbour at the time instance, then the context vector for the hardwired attention model is defined as,

The simplest repesentation scheme can be given by,

where is the distance between the neighbour and the pedestrian of interest at the time instance.

We then employ a simple concatenation layer to combine the information from individual attentions. Hence the combined context vector can be denoted as,

where is refered to as the set of weights for concatenation.We learn this weight value also through back-propagation.

The final prediction can be computed as,

Model learning

The given input trajectories in the training set are clustered used DBSCAN based on source and sink positions and we run an outlier detection algorithm for each cluster considering the entire trajectory.

In the training phase trajectories are clustered based on the entire trajectory and after clustering we learnt a separate trajectory prediction model for each generated cluster. When modelling the local neighbourhood of the pedestrians of interest, we have encoded the trajectories of those closest 10 neighbours in each direction, namely 165 front, left and right.

To select the appropriate prediction model to use, the mean trajectory for each cluster for the period of to is generated and in the testing phase, the given trajectories are assigned to the closest cluster centre while considering those mean trajectories as the cluster centroids.

Comparison to the Social-LSTM model

  1. The proposed combined attention model considers the entire sequence of hidden states for both the pedestrian of interest and his or her neighbours
  2. Then we utilise time dependent weights which enables us to vary their influence in a timely manner.
  3. Learning a different trajectory prediction model for each trajectory cluster.

Experiments

image_1c7e9jpbu18sk1nc1bj81a85bh147.png-115.9kB
For the SF model, preferred speed, destination, and social grouping factors are used to model the agent behaviour.

For S-LSTM model a local neighbourhood of size 32px was considered and the hyper-parameters were set according to Alahi et al(2016).

In order to evaluate the strengths of the proposed model, we compare this combined attention model (OURcmb) and two variations on our proposed approach: 1) OURsft, 265 which ignore the neighbouring trajectories and considers only the soft attention component derived from the trajectory of the person of interest when making predictions; and 2) OURsc which omits the clustering stage such that only a single model (using combined attention weights) is learnt

添加新批注
在作者公开此批注前,只有你和作者可见。
回复批注