Skip to content

Flight Load Factor Predictions based on Analysis of Ticket Prices and other Factors

The ability to forecast traffic and to size the operation accordingly is a determining factor, for airports. However, to realise its full potential, it needs to be considered as part of a holistic approach, closely linked to airport planning and operations. To ensure airport resources are used efficiently, accurate information about passenger numbers and their effects on the operation is essential. Therefore, this study explores machine learning capabilities enabling predictions of aircraft load factors. The rationale behind the logic used stems from the assumption that using past traffic statistics in a form of historic load factor may not be sufficient, especially at times of high traffic volatility such as during regional bank holidays. Therefore, exploration efforts were made to parameterize some novel predictive elements that could provide passenger demand predictions at different granularity levels. The investigation has been successful and through the use of gradient boosting technique, the model, including 9 significant predictors was created. The load factor predictions per flight perform highly accurately with an average mean absolute error around 10 percentage points. In principle, this achievement outscores any other related work conducted in this domain to date. On top of that, the model itself is scalable and can be applied to any airport in the network as applied to use cases within the presented paper.

Miroslav Spak, Lorenzo Frigerio, Lenka Hanakova, Vladimir Socha, Rocio Barragan Montes, Vincent Treve, Flight Load Factor Predictions based on Analysis
of Ticket Prices and other Factors, In Proc. of the SESAR Innovation Days, December 2025.

Click here to access the paper.

Releated Posts

Investigating a Feature Unlearning Bias Mitigation Technique for Cancer-type Bias in AutoPet Dataset

We proposed a feature unlearning technique to reduce cancer-type bias, which improved segmentation accuracy while promoting fairness across sub-groups, even with limited data.
Read More

Muppet: A Modular and Constructive Decomposition for Perturbation-based Explanation Methods

The topic of explainable AI has recently received attention driven by a growing awareness of the need for transparent and accountable AI. In this paper, we propose a novel methodology to decompose any state-of-the-art perturbation-based explainability approach into four blocks. In addition, we provide Muppet: an open-source Python library for explainable AI.
Read More