From Hawkes Processes to Attention: Time-Modulated Mechanisms for Event Sequences
By: Xinzi Tan , Kejian Zhang , Junhan Yu and more
Marked Temporal Point Processes (MTPPs) arise naturally in medical, social, commercial, and financial domains. However, existing Transformer-based methods mostly inject temporal information only via positional encodings, relying on shared or parametric decay structures, which limits their ability to capture heterogeneous and type-specific temporal effects. Inspired by this observation, we derive a novel attention operator called Hawkes Attention from the multivariate Hawkes process theory for MTPP, using learnable per-type neural kernels to modulate query, key and value projections, thereby replacing the corresponding parts in the traditional attention. Benefited from the design, Hawkes Attention unifies event timing and content interaction, learning both the time-relevant behavior and type-specific excitation patterns from the data. The experimental results show that our method achieves better performance compared to the baselines. In addition to the general MTPP, our attention mechanism can also be easily applied to specific temporal structures, such as time series forecasting.
Similar Papers
Hyper Hawkes Processes: Interpretable Models of Marked Temporal Point Processes
Machine Learning (Stat)
Lets computers predict events and explain why.
Advances in Temporal Point Processes: Bayesian, Neural, and LLM Approaches
Machine Learning (CS)
Helps predict when things will happen next.
Differentiable Adversarial Attacks for Marked Temporal Point Processes
Machine Learning (CS)
Tricks computer models that track events.