On the Simplification of Neural Network Architectures for Predictive Process Monitoring
By: Amaan Ansari, Lukas Kirchdorfer, Raheleh Hadian
Potential Business Impact:
Makes smart computer predictions much faster and smaller.
Predictive Process Monitoring (PPM) aims to forecast the future behavior of ongoing process instances using historical event data, enabling proactive decision-making. While recent advances rely heavily on deep learning models such as LSTMs and Transformers, their high computational cost hinders practical adoption. Prior work has explored data reduction techniques and alternative feature encodings, but the effect of simplifying model architectures themselves remains underexplored. In this paper, we analyze how reducing model complexity, both in terms of parameter count and architectural depth, impacts predictive performance, using two established PPM approaches. Across five diverse event logs, we show that shrinking the Transformer model by 85% results in only a 2-3% drop in performance across various PPM tasks, while the LSTM proves slightly more sensitive, particularly for waiting time prediction. Overall, our findings suggest that substantial model simplification can preserve predictive accuracy, paving the way for more efficient and scalable PPM solutions.
Similar Papers
Directly Follows Graphs Go Predictive Process Monitoring With Graph Neural Networks
Machine Learning (CS)
Helps computers understand messy business steps better.
Working My Way Back to You: Resource-Centric Next-Activity Prediction
Machine Learning (CS)
Helps predict who does what next.
From Source to Target: Leveraging Transfer Learning for Predictive Process Monitoring in Organizations
Machine Learning (CS)
Lets companies predict problems using other companies' data.