论文标题
主动:活动序列的自我训练的时间点过程流动
ProActive: Self-Attentive Temporal Point Process Flows for Activity Sequences
论文作者
论文摘要
任何人类活动都可以表示为实现某个目标的行动的时间顺序。与机器制造的时间序列不同,这些动作序列是高度分散的,因为完成类似的动作所花费的时间可能会有所不同。因此,了解这些序列的动力学对于许多下游任务,例如活动长度预测,目标预测等。建模活动序列的现有神经方法限于视觉数据,或者是特定于任务的,即仅限于下一个动作或目标预测。在本文中,我们提出了积极主动的,这是一个神经标记的时间点过程(MTPP)框架,用于建模活动序列中动作的连续时间分布,同时解决三个高影响力问题 - 下一个动作预测,序列目标预测和端到端序列。具体而言,我们利用一个具有时间归一化流量的自我发项模块来对序列中的影响和到达的影响。此外,对于时间敏感的预测,我们通过基于边缘的优化程序进行了序列目标的早期检测。这种往返允许积极主动使用有限数量的动作来预测序列目标。从三个活动识别数据集得出的序列进行的广泛实验表明,就行动和目标预测而言,主动的准确性提升了,并且是端到端动作序列生成的首次应用。
Any human activity can be represented as a temporal sequence of actions performed to achieve a certain goal. Unlike machine-made time series, these action sequences are highly disparate as the time taken to finish a similar action might vary between different persons. Therefore, understanding the dynamics of these sequences is essential for many downstream tasks such as activity length prediction, goal prediction, etc. Existing neural approaches that model an activity sequence are either limited to visual data or are task specific, i.e., limited to next action or goal prediction. In this paper, we present ProActive, a neural marked temporal point process (MTPP) framework for modeling the continuous-time distribution of actions in an activity sequence while simultaneously addressing three high-impact problems -- next action prediction, sequence-goal prediction, and end-to-end sequence generation. Specifically, we utilize a self-attention module with temporal normalizing flows to model the influence and the inter-arrival times between actions in a sequence. Moreover, for time-sensitive prediction, we perform an early detection of sequence goal via a constrained margin-based optimization procedure. This in-turn allows ProActive to predict the sequence goal using a limited number of actions. Extensive experiments on sequences derived from three activity recognition datasets show the significant accuracy boost of ProActive over the state-of-the-art in terms of action and goal prediction, and the first-ever application of end-to-end action sequence generation.