7.9
CiteScore
 
3.6
Impact Factor
Generic selectors
Exact matches only
Search in title
Search in content
Post Type Selectors
Search in posts
Search in pages
Filter by Categories
ABUNDANCE ESTIMATION IN AN ARID ENVIRONMENT
Case Study
Correspondence
Corrigendum
Editorial
Full Length Article
Invited review
Letter to the Editor
Original Article
Research Article
Retraction notice
REVIEW
Review Article
SHORT COMMUNICATION
Short review
7.2
CiteScore
3.7
Impact Factor
Generic selectors
Exact matches only
Search in title
Search in content
Post Type Selectors
Search in posts
Search in pages
Filter by Categories
ABUNDANCE ESTIMATION IN AN ARID ENVIRONMENT
Case Study
Correspondence
Corrigendum
Editorial
Full Length Article
Invited review
Letter to the Editor
Original Article
Research Article
Retraction notice
REVIEW
Review Article
SHORT COMMUNICATION
Short review
View/Download PDF

Translate this page into:

Research Article
ARTICLE IN PRESS
doi:
10.25259/JKSUS_1553_2025

Seasonal inflation dynamics: Insights from probabilistic and hybrid forecasting models

Department of Statistics, The Islamia University of Bahawalpur, Punjab, 63100, Pakistan
Department of Mathematics and Statistics, College of Science, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh, 11432, Saudi Arabia
Department of Statistics, Faculty of science, University of Tabuk, Tabuk, 47512, Saudi Arabia
Department of Statistics, Mathematics, and Insurance, Benha University, Benha 13511, Egypt

* Corresponding author: E-mail address: iielbatal@imamu.edu.sa (I Elbatal)

Licence
This is an open-access article distributed under the terms of the Creative Commons Attribution-Non Commercial-Share Alike 4.0 License, which allows others to remix, transform, and build upon the work non-commercially, as long as the author is credited and the new creations are licensed under the identical terms.

Abstract

Inflation significantly influences economic growth and policy decisions, making accurate forecasting essential. This study analyzes seasonal inflation patterns using monthly data covering the period 2003–2024. It also evaluates a wide range of forecasting methods, including autoregressive integrated moving average (ARIMA), artificial neural networks (ANNs), exponential smoothing techniques (Brown and Holt–Winters), BATS (Box-Cox transformation ARMA errorsTrend), TBATS (Trigonometric seasonality Box-Cox transformation ARMA errorsTrend, and Seasonal components), and hybrid models, in addition to probabilistic distributions such as the Weibull, exponentiated Weibull, Kumaraswamy Weibull, exponential, and inverse Weibull. Results show that ANNs achieved the highest predictive accuracy, while the Weibull distribution best captured seasonal dynamics. The 10-month forecast horizon, extending from June 2024 to March 2025, reveals a stable yet gradually rising inflation pattern. The findings demonstrate the superior forecasting accuracy of ANN-based models and underscore the study’s unique integration of machine-learning techniques with probabilistic models to improve inflation-forecasting performance and support informed economic policy.

Keywords

Data analysis
Inflation forecasting
Maximum likelihood
Time series
Weibull distribution

1. Introduction

Inflation is one of the most challenging and complex processes to define and quantify (Baciu, 2015). It is a process where money depreciates and the price index increases. It is a steady rise in the economy’s overall price level. Gross domestic product, money supply, exports, import prices, exchange rates, interest rates, fiscal policy, and more are all intimately correlated with the rate of inflation (Bashir et al., 2011). Significant shifts in the economy’s supply and demand can result in volatile inflation. In addition to making long-term planning challenging, the erratic and unexpected inflation rates deter both savings and investment (Baciu, 2015). The unpredictability of inflation has two different kinds of economic repercussions. The first impacts businesses and customers who have to make financial decisions based on projected future inflation. According to (Devereux, 1989), the uncertainty surrounding inflation in this situation influences financial markets by increasing interest rates, creating uncertainty about other factors crucial in economic decisions, and motivating businesses to allocate their resources to mitigate specific inflation-related risks. The second consequence occurs when the inflation rate deviates considerably from the forecast. A transfer of wealth from the creditor to the debtor occurs when inflation is better than anticipated. Because repayment is made with funds whose purchasing power is penalized by the inflation rate, the latter is preferred. Some contend that because inflation affects both the economy and society, it is not a desirable state of affairs (Khan and Gill, 2010). Both monetary and fiscal views were unable to definitively identify the causes of the rising inflation. Real income is randomly redistributed as a result of inflation’s effects on income distribution. A nation’s financial situation deteriorates when inflation raises the level of prices for goods, services, and other items. Therefore, it can be concluded that one of the main causes of rising commodity prices in every economy is inflation. The state of the economy is mostly determined by inflation (Suseno, 2009), thus it requires careful consideration from a variety of sources, particularly the monetary authorities in charge of managing inflation. A nation’s economic system and businesspeople in general cannot undervalue the rate of inflation. Annual inflation rate forecasts are very important for monetary policy as well as for planning in a variety of corporate and economic scenarios. Several studies have looked into ways to make inflation predictions better. For instance, (Stock and Watson, 1999) enhance their inflation projections by utilizing data from other economic indicators. Disaggregated data is used by (Hubrich, 2005). Additionally, (Camba and Kapetanios, 2004), (Hofmann, 2009), and (Angelini, et al., 2001) consider gleaning valuable data for inflation rate forecasts from huge datasets. Numerous studies have examined the topic of forecasting seasonal time series; (Franses, 1991; Paap et al., 1997), and for a more general discussion of seasonality (Hylleberg, 2014). A summary of the findings is given by (Osborn 2004). Forecasting is a technique that uses historical data to estimate future values. Among the time series data is information on inflation. It is possible to forecast future time data by modelling historical time data. In this study, we investigate the use of time series and probability models to anticipate inflation in the United States. To examine US inflation and short-term projections, we contrasted hybrid and non-hybrid models of time series and the probability distributions.

The majority of recent research on inflation forecasting still uses conventional econometric techniques like vector autoregression (VAR) and autoregressive integrated moving average (ARIMA) models, despite the wealth of prior research in this area. It is challenging for these conventional models, which often assume linear relationships, to explain for the seasonal and nonlinear variations commonly observed in real-world data. The complex patterns and anomalies that have influenced the historical evolution of inflation are, therefore, sometimes difficult for individuals to comprehend. It is also not often recognized how probabilistic and hybrid forecasting techniques might be used, particularly when studying market-driven seasonal inflation trends: expectations, cyclical economic swings, and changes in monetary policy. Prior studies have mostly examined quarterly or annual data, neglecting the short-term seasonal variations that may serve as a precursor to inflationary trends. The relative usefulness of complex hybrid models, such as ARIMA-ANN or ARIMA-TBATS, in forecasting inflation has also not been thoroughly examined in many studies; as a result, nothing is known here. Examine its performance by contrasting it with other machine learning methods, including ANN.

Additionally, there is not enough empirical data to evaluate the effects of combining learning-based and probabilistic models on the long-term stability and comprehensibility of inflation forecasts. By using both probabilistic and hybrid forecasting models to investigate seasonal inflation dynamics, the current study closes these gaps. This method increases the accuracy and adaptability of inflation forecasts while also bridging the gap between traditional statistical methods and modern machine learning techniques. Examining such complex modeling frameworks is essential since inflation prediction plays a major role in monetary policy, economic planning, and investment strategies. By identifying models that effectively capture seasonal and nonlinear inflation patterns, the work provides significant insights for researchers and policymakers seeking more reliable forecasting tools.

2. Objectives of the Study

This study’s primary objective is to apply probabilistic and hybrid forecasting techniques to analyze and forecast seasonal inflation patterns. The study’s goal is to identify the model that best captures the nonlinear, stochastic, and seasonal tendencies seen in inflation data.

The specific objectives of the study are as follows:

  • Examine the seasonal variations and structural characteristics of the breakeven inflation rate using advanced time series analysis.

  • Using probabilistic models to show the stochastic and probabilistic nature of inflation data, such as the inverse Weibull (IW), Weibull (W), exponentiated Weibull (EW), Kumaraswamy Weibull (KW), and exponential (E) models.

  • Hybrid forecasting models such as ARIMA-TBATS, ARIMA-error, trend and seasonality (ARIMA-ETS), ARIMA-ANN, and error, trend and seasonality-artificial neural network-trigonometric seasonality Box-Cox transformation ARMA errors trend, and seasonal components (ARIMA-ETS-ANN-TBATS-STLM) should be developed and evaluated to improve prediction accuracy.

  • To ascertain which method best captures seasonal and nonlinear inflation trends, the effectiveness of probabilistic, hybrid, and pure machine learning models is assessed.

  • The anticipated accuracy of each model is evaluated using statistical performance measures such as mean absolute error (MAE), mean absolute percentage error (MAPE), mean absolute scaled error (MASE), and root mean square error (RMSE).

  • To provide decision-makers and economists with relevant data that will enhance their understanding of inflation dynamics and facilitate the development of more astute, fact-based policies.

3. Data

The Federal Reserve Bank of St. Louis (United States) provided the 10-year breakeven inflation rate (T10YIEM) data used in this study. Evaluation of the data is available to the public at the repository at https://fred.stlouisfed.org/series/T10YIEM. Our data are seasonal, with a total of 257 observations. This dataset is categorized as a univariate time series, meaning that the only significant component is the 10-year breakeven inflation rate. This indicator provides crucial information about long-term pricing trends and is based on market-based forecasts of future inflation. The study focused on evaluating the prediction validity of several individual and hybrid time series models rather than incorporating any exogenous or explanatory elements. The R programming language was used for time series analysis and probability models. The models’ accuracy in representing and forecasting changes in inflation situations in the USA over time was assessed.

3.1 Pre-processing steps of the dataset

The data went through a number of processing stages before modeling to make sure it was suitable for analysis. After the dataset was examined for consistency and completeness, any missing values were filled in using linear interpolation to preserve continuity. Then, beginning in January 2003, the inflation series was transformed into a monthly time series object with a frequency of 12 to maintain its seasonal and temporal structure. Trends, seasonal patterns, and the necessity of differencing in models like ARIMA were identified by preliminary visualization and stationarity tests, such as the Augmented Dickey-Fuller (ADF) test.

Fig. 1 provides scatter plots for Lag-1 and Lag-2 values, with points color-coded by month. The observed linear pattern suggests strong temporal dependence, indicating that past inflation values significantly influence future values. The color-coded distribution further highlights seasonal variations, emphasizing that inflation trends exhibit distinct patterns across different months. In addition to time series analysis, the study employs visual tools to understand data distribution and variability. Additionally, Fig. 1 shows that a strong positive linear relationship between successive values in the time series is indicated by the lag plot, which displays yt versus yt1 .

(a) Lag 1 and lag 2 Scatter plots and (b) ACF and PACF plots of series.
Fig. 1.
(a) Lag 1 and lag 2 Scatter plots and (b) ACF and PACF plots of series.

Fig. 2 shows that this plot is a combination of a histogram and a kernel density estimate (KDE) for the variable y and the total time on test (TTT) plot; the descriptive statistics are completed. The histogram is in black bars, and the KDE is in a blue line. The shape of the density is left-skewed, and the TTT plot shows that the data are in an increasing trend.

(a) Histogram with kernel density and (b) TTT plot
Fig. 2.
(a) Histogram with kernel density and (b) TTT plot

A graphical approach to evaluate whether the failure rate of a dataset (often referred to as its hazard rate) is increasing, decreasing, or stable over time is represented by the TTT graph. Prior to selecting a statistical model (such as lognormal, Weibull, or exponential), it is beneficial to observe the fundamental pattern of life distribution.

Fig. 3 presents a violin plot, which combines density estimation with a box plot to display the distribution of inflation data. The black bar represents the interquartile range (IQR), and the white dot indicates the median. The shape of the violin plot suggests that while the central part of the data is densely packed, the distribution widens significantly toward the tails, reflecting variations in inflation trends over time. Additionally, Fig. 3 uses a boxplot to summarize key statistical properties of inflation data, illustrating the minimum, first quartile (Q1), median, third quartile (Q3), and maximum values. The presence of whiskers and potential outliers helps in identifying seasonal variations and extreme fluctuations in inflation trends.

(a) Violin plot and (b) box plot
Fig. 3.
(a) Violin plot and (b) box plot

Table 1 shows the dataset’s intermediate value, where half of the data is above, and half is below. The value of the first quartile is 1.830, the third quartile is 2.360, and the minimum value is 1.06. 2.008 is the average of all the values in the dataset. 2.88 is the greatest value in the dataset. The information explains the data’s range, central tendency, and spread.

Table 1. Basic statistical measures of the seasonal inflation percentage in the United States.
Min 1st Qu Median Mean 3rd Qu Max
1.06 1.830 2.19 2.088 2.360 2.88

4. Conceptual Framework And Definition of Variable

Within the conceptual framework, independent and dependent variables serve as the foundation for analyzing potential cause–and–effect relationships. As shown in Fig. 4, the independent variables considered are money supply (M3), exchange rate (ER), and gross domestic product (GDP). The dependent variable, represented as a percentage, is the inflation rate (IR).

The model’s conceptual framework for the variables.
Fig. 4.
The model’s conceptual framework for the variables.

The study’s conceptual underpinning is the idea that significant macroeconomic variables, such as the M3, ER, and GDP, affect inflation. The seasonal and directional tendencies of inflation are influenced by the interaction of various elements across time. It was hoped that this theoretical framework would be converted into quantitative models in order to reflect these processes through empirical investigation. Most accurate results were generated by the ANN model, even though a range of hybrid and probabilistic forecasting techniques were used. The fundamental idea is that traditional linear or somewhat mixed models are unable to adequately represent the complex and nonlinear character of inflation dynamics. This study supports the idea that representation should be egalitarian.

The theoretical hypothesis that nonlinear modelling techniques can better capture the seasonal and structural complexity of inflation trends is thus supported by the improved forecasting capabilities of the ANN model, which demonstrate the relationship between the conceptual framework and the empirical data.

5. Modeling Methodology

Time series forecasting techniques included ARIMA, Brown, Holt, Winters, ANN, ETS, ARFIMA, BATS, TBATS, Box-Cox parameter, Box-Cox transformation, Croston, structural, mean forecast, exponential smoothing forecasts, Theta, linear, and Bagged. These methods were chosen because of their efficacy in time series analysis. A range of forecasting techniques is used to analyze the trends of seasonal inflation, and the relative benefits of statistical, machine learning, and hybrid approaches are assessed.

5.1 Auto regressive integrated moving average

ARIMA was chosen due to its ability to accurately capture linear time-dependent patterns and short-term autocorrelation in inflation data. It offers a simple framework for comparing more intricate methods.

The methodology of time series modelling and forecasting consists of gaining meaningful measures of statistics and other properties of the time series data. The ARIMA model can be shown as ARIMA(p,d,q), where, p indicates the order of AR, d shows the differenced trend, and q shows the moving average order. Time series models can be shown as in Eq. (1) and Eq. (2), respectively (Franses, 1991).

(1)
Yt= φ1 Yt1 + φ2 Yt2 ++ φp Ytp +εt

and

(2)
 Yt= θ1 εt1  θ2 εt2  θq εtq +εt.

The ARMA model can be generally in the form of Eq. (3);

(3)
Yt= α+φ1 Yt1 + φ2 Yt2 ++ φp Ytp +εt θ1 εt1  θ2 εt2  θq εtq .

5.2 Exponential smoothing model (ESM)

The ESM is widely employed for time series forecasting, particularly in short-term applications (Gardner, 1985). It is based on the principle that future values of a variable are influenced by its past observations, with greater weight assigned to more recent data. The method applies an exponentially declining weighting scheme across successive observations. The general mathematical form of the ESM is expressed as:

(4)
f st = αxt+ 1α st1 .

After simplification, Eq. (4) results in

(5)
f st =st1 +α xtst1 .

Here the

st=Smoothed statistic or the weighted average of the current observation xt

st1 = Onetime lagged smoothed statistic .

α= Smoothingparameterrangesfrom  0<α<1.

xt = Current period.

Exponential smoothing is a commonly applied technique in time series forecasting, where past observations are weighted with exponentially decreasing importance over time. The fundamental principle that current observations are more important for making predictions than earlier observations is captured by giving more weight to recent data. It does very well with data that shows trends and seasonality. Both seasonal and systematic patterns can be present in data when using exponential smoothing, or ETS. It is an important forecasting approach that can be used in place of ARIMA methods.

The ability of these exponential smoothing techniques to accurately represent the level, trend, and seasonal characteristics of time series data led to their selection. The Brown model analyzes seasonal inflation using trend smoothing, the Holt technique considers both trend and level, and the Winters model takes seasonal variations into account.

5.3 Artificial neural networks

The ANN has garnered a lot of interest in a variety of scientific and technical domains. They belong to a class of non-linear models that can learn from data and are modeled after the structure of the human brain. In domains where, conventional statistical techniques were previously prevalent, ANNs have been used for tasks like pattern recognition, classification, prediction, and process control.

ANNs are useful from a statistical standpoint because they offer tractable multivariate non-linear approaches, and statistical science provides theoretical underpinnings that support the advancement of neurocomputing. Because of this intersection, ANN is very attractive for issues like categorization and prediction. They have been used, for instance, to predict stock market performance, detect undersea sonar connections, and diagnose cardiac illness in patients. ANN is one of the most used techniques in bioinformatics for predicting secondary structures of proteins.

ANNs are employed to model intricate and nonlinear interactions that traditional statistical methods may overlook. Its remarkable learning ability and flexibility enable it to uncover concealed patterns in inflation data. The data indicate that the ANN model surpassed other methods, showcasing its predictive strength and robustness.

5.4 Error, trend, and seasonality (ETS)

In the ETS model, the three letters correspond to error (E), trend (T), and seasonality (S). The first symbol specifies the error type (“A” for additive, “M” for multiplicative, or “Z” for automatic selection), the second indicates the trend component (“N” for none, “A” for additive, “M” for multiplicative, or “Z”), and the third denotes the seasonal component with the same options. For instance, the multiplicative Holt–Winters method with multiplicative error is written as ETS (M, A, M), while simple exponential smoothing with additive error is denoted ETS (A, N, N). The model can be estimated in R using the ETS function from the forecast package, which automatically selects the most appropriate specification and estimates the parameters.

These exponential smoothing techniques were selected due to their capacity to accurately represent the seasonal, trend, and level characteristics of time series data. The Brown model uses trend smoothing to examine seasonal inflation, the Holt technique considers both trend and level, and the Winters model takes seasonal changes into account.

5.5 Autoregressive fractionally integrated moving average (ARFIMA)

Time series data can be modeled using the ARFIMA model. Unlike the standard ARIMA model, ARFIMA allows the differencing parameter to take non-integer values, enabling it to capture long-range dependence in data. The ARFIMA (p, d, q) specification is a well-known member of the long-memory model family.

ARFIMA was used to analyze the long-memory characteristics and persistence of inflation data, as well as the potential fractional differencing behavior that is commonly seen in macroeconomic time series.

5.6 BATS

The BATS model works well with time series data that exhibit variable variance over time and complex or varied seasonality. Specifically, compared to simpler models like ARIMA, it can manage a variety of seasonal impacts and non-linearities more readily. A class of time series forecasting models is referred to by the initials BATS, which stands for Box-Cox transformation, ARMA errors, Trend, and Seasonal components. Because it is designed to handle complicated seasonal patterns and non-linearities in time series data, it is particularly useful for series with several seasonal periods, heteroscedasticity (changing variance), and nonlinearity.

The BATS and TBATS models are chosen because of their ability to represent complex periodic fluctuations, nonlinear patterns, and a wide range of seasonality that are difficult for conventional models to capture. For seasonal data at high frequency, TBATS specifically enhances adaptation by integrating ARMA errors and Box-Cox transformation.

5.7 TBATS

A non-linear time series model called TBATS can manage data series with several seasonal patterns, meaning that the data’s pattern varies with time. Because Trigonometric Seasonality (TBATS) can handle complex and high frequencies, this approach is recommended over BATS.

To apply TBATS models to the data series, the “tbats ()” function was utilized.

5.8. Theta model

The theta model is a forecasting technique that estimates future values by decomposing a time series into modified versions, known as theta lines. It captures the average change between successive observations and then extrapolates this pattern to generate forecasts. The model is mathematically expressed as follows:

pd+1 = 1/m pd + pd1 +pdm1 .

The Theta model was chosen because of its outstanding results in forecasting contests involving macroeconomic time series, such as the M3 competition. To capture both short-term and long-term fluctuations, it divides the original series into “theta lines” using a trend correction technique in combination with basic exponential smoothing. For this reason, it works well for inflation data that has both trend and seasonal features.

5.9 Croston’s method

A forecasting technique called the Croston model was created for patterns of intermittent demand, which are defined by extended stretches of zero demand interspersed with non-zero demand periods. It operates by splitting the sporadic time series into two distinct series: one for the values of the non-zero demands and another for the time intervals between them. Forecasts are then produced by applying simple exponential smoothing to both of these new series.

Croston’s approach is often used in inventory or demand forecasting, and it can also be assessed for its applicability to inflation data, which makes a comprehensive model comparison easier. The limits of traditional smoothing methods when working with data that has an uneven distribution are highlighted by this method.

5.10 Structural

A structural time series model, which frequently makes use of a state space representation, breaks down a time series into significant, interpretable elements such as trend, seasonality, and irregular fluctuations. By extending these components, these models are useful for forecasting and comprehending the underlying dynamics of the data.

The capacity of the state-space-based structural time series model to break down a time series into distinct elements, such as trend, seasonality, and random fluctuations, led to its selection. This method works particularly well in two areas: analyzing inflation data to identify underlying patterns and determining how important each element is in shaping overall patterns. It provides an interpretable and flexible framework for identifying deterministic and stochastic trends in economic data.

5.11 Mean forecast

The mean forecast model in time series analysis is basically a random forecast model that has been modified especially for time series data forecasting or classification applications. In time series data, sequential dependencies are not automatically taken into consideration by random forecasts. To use them successfully, then, changes are required. To maintain the chronological order of data during training and validation, these changes frequently entail developing lag variables (using historical values as predictors).

The mean forecast model, a nonlinear ensemble forecasting method that leverages decision tree capabilities and uses averaging to enhance generalization, was selected. To reduce variance and increase forecast reliability, the mean forecast model integrates forecasts from many decision trees rather than relying on a single decision tree, which may be unstable due to its sensitivity to slight changes in the data. As a result of its use in the study, it is possible to assess how well tree-based ensemble approaches capture the complex and nonlinear interactions present in inflation dynamics in contrast to conventional statistical and hybrid models. Despite the mean forecast model’s initial design for regression and classification, its application to time series data offers important insights into how well it handles economic data with irregular patterns and structural fluctuation.

5.12 Bagged model

As an ensemble forecasting technique that reduces variance and improves model stability, the Bagged model was included. Bagging lowers overfitting and increases prediction accuracy by averaging predictions from several resampled data sets. Integrating allows one to assess ensemble-based averaging’s resilience against each prediction model.

5.13 Linear model

The linear model, a basic statistical technique, was used to assess if inflation dynamics could be sufficiently captured by a straightforward linear relationship. Its inclusion offers a benchmark for evaluating the added value that more intricate hybrid and nonlinear models bring. In spite of its simplicity, it helps determine the extent to which sophisticated modeling methods improve prediction accuracy.

5.14 ARIMA-ANN

A hybrid ARIMA-ANN model was utilized in this investigation to identify both linear and nonlinear patterns in the time series data. Two consecutive steps were taken in the forecasting procedure. Initially, the original time series’ linear components were modeled using an ARIMA model. Analyzing the autocorrelation function (ACF) and partial autocorrelation function (PACF) plots allowed us to ascertain the orders of the ARIMA model. Following that, an ANN was fed the residuals from the ARIMA model, which include the nonlinear structure that ARIMA was unable to capture. Through trial and error and cross-validation, the ANN parameters, such as the number of hidden layers, neurons, activation function, and learning rate, were adjusted to reduce the accuracy metrics. Ultimately, the ARIMA prediction was added to the ANN prediction of the residuals to create the hybrid forecast. This methodical technique guarantees that the ANN component represents the remaining nonlinear structures, increasing overall forecasting accuracy, while the ARIMA component efficiently catches linear trends.

5.15 ARIMA-ETS hybrid model

ARIMA was initially used to model the linear components of the time series in the ARIMA-ETS hybrid model. The ARIMA residuals were then subjected to the ETS model in order to capture more seasonal and trend structures, especially those that exhibit gradual changes over time. The ARIMA and ETS forecasts were combined to create the final forecast, which enabled the model to take into consideration both linear and smooth nonlinear patterns.

5.16 ARIMA–TBATS hybrid model

Once more, the linear components of the series were modelled using ARIMA in the ARIMA–TBATS hybrid model. After that, TBATS was used on the ARIMA residuals to deal with any lingering nonlinear trends and complex seasonal patterns, such as multiple or non-integer seasonal cycles. Both linear structures and complex seasonal changes were successfully represented by combining the results of the ARIMA and TBATS forecasts into the final prediction.

5.17 ARIMA–ETS–ANN–TBATS–STLM

To capture distinct patterns in the data, the ARIMA–ETS–ANN–TBATS–STLM hybrid model used each component in turn: ANN for the remaining nonlinear structure, ETS and TBATS for overall trend and seasonality, and ARIMA for linear trends. To address local seasonal patterns, STLM was incorporated into the hybrid architecture. The outputs of all the components were combined to create the final forecast.

5.18 Accuracy measures

To evaluate the performance of each forecasting model, we used the following accuracy measures: RMSE, mean absolute error (MAE), and MAPE. A model demonstrates better predictive performance when the values of RMSE, MAE, MAPE, and MASE are smaller.

The RMSE calculates the difference between observed and predicted values, and it is given by:

RMSE= 1n i=1 n viv^i 2 .

The average of all absolute errors is known as MAE. It is defined by:

MAE= 1n i=1 n|viv^i|.

The MAPE quantifies the average percentage deviation between forecasted and actual values. It is defined as:

MAPE= 1n i=1 n viv^i vi.

The accuracy of forecasts can be measured by the MASE. It was designed to be scale-independent and compares the forecast error to the error from a baseline model. The MASE is defined by:

MASEv=mean |ej| 1 T1 t=2 T|vtvt1 |.

6. Results and Discussion

Four common accuracy metrics were used to evaluate the forecasting ability of several time series and machine learning models for monthly inflation in the United States: RMSE, MAE, MAPE, and MASE. Table 2 presents the fitted models along with their forecasting accuracy measures for monthly inflation in the USA. All of these indicators show lower values for higher anticipated accuracy and model reliability. By far the most accurate predictions were produced by the ANN of all the models that were examined. The ANN’s remarkable capacity to capture the nonlinear, dynamic, and interactive nature of inflation, a process influenced by a variety of factors like shifts in energy costs, labor market conditions, global supply disruptions, and monetary policy actions, is demonstrated by this result. By dynamically learning complex connections within data, the ANN adjusts to structural and regime changes in the economy, unlike classic linear techniques that are predicated on set assumptions. With the lowest RMSE (0.1132), MAE (0.0809), MAPE (4.80%), and MASE (0.2298), the ANN model performs better than all other models. It is about 10–11% better than ARIMA and 5–7% better than hybrid models like ARIMA-ANN and ARIMA-ETS-ANN-TBATS. ANN offers the most reliable and accurate short-term inflation forecast, as seen by the larger errors of traditional models such as ARIMA, Holt-Winters, ETS, and ARFIMA.

Table 2. Fitted models along with their forecasting accuracy measures for monthly inflation in the USA.
Models RMSE MAE MAPE MASE
ANN 0.1132 0.0809 4.7993 0.2298
ARIMA-ANN 0.1176 0.0845 5.0792 -
ARIMA-ETS-ANN-TBATS-STLM 0.1196 0.0842 5.0503 -
ARFIMA 0.1236 0.0899 5.4249 0.2555
ARIMA 0.1268 0.0901 5.3815 0.2559
Holt-winters 0.1268 0.0901 5.3815 0.2559
ETS 0.1339 0.0946 5.5146 0.2689
Structural 0.1365 0.0951 5.8852 0.2703
Theta 0.1366 0.0951 5.8855 0.2703
BATS 0.1247 0.0892 5.5758 0.9335
TBATS 0.1247 0.0892 5.5758 0.9335
Bagged 0.1668 0.1131 7.9758 1.1844
Linear 0.3956 0.3099 20.22 0.8804
Mean forecast 0.3995 0.3164 20.4622 0.8988
Crostons 0.3102 0.2168 14.8386 0.6160
Box-Cox transformation 0.2334 0.1762 52.8963 0.2595
Box cox parmeter 0.2059 0.11707 9.0809 0.3326
ARIMA-TBATS 0.1241 0.0891 5.4532 -
ARIMA-ETS 0.1294 0.0917 5.4113 -

The ARIMA–ANN hybrid model achieved the second-best results, demonstrating the advantages of combining the nonlinear learning capabilities of ANN with the linear forecasting framework of ARIMA. In a similar vein, the combination of ARIMA, ETS, ANN, TBATS, and STLM showed competitive performance, indicating that hybrid frameworks can successfully include complementary model strengths to simultaneously capture abnormal fluctuations, trends, and seasonality. These models can handle both short-term volatility and long-term persistence in inflation behavior due to the combination of machine learning and statistics.

Traditional techniques like ARIMA, ARFIMA, and Holt-Winters have a moderate but consistent accuracy, which reflects their ability to identify patterns and seasonal components, but their limited ability to adjust to abrupt changes in the economy. As inflation volatility rose, models with somewhat bigger errors, like ETS, structural, and theta, showed less response. Overfitting compromises a model’s out-of-sample performance by fitting the previous data too closely. Despite successfully capturing intricate seasonal patterns in this investigation, the BATS and TBATS models’ comparatively high MASE values raise the possibility of overfitting. Due to component redundancy and model complexity, the Bagged ensemble model also showed good in-sample performance but poorer generalizability.

Simpler methods like mean forecast, Croston’s model, and linear regression provide noticeably higher error levels, which highlights their inadequacies in capturing the dynamic and feedback-driven mechanisms of inflation. The poor performance of the Box–Cox transformation model shows that stabilizing variance alone is insufficient to explain asymmetric and nonlinear reactions to inflation. The computed transformation parameter (λ = 0.2059) provided no variance adjustment and had low predictive power.

The models with the lowest accuracy in simulating inflation were the Box-Cox transformation, Croston’s method, and linear regression. Because the Box–Cox method concentrates on variance stability rather than temporal dependence, it fails when inflation is affected by asymmetric or persistent shocks. Assuming that events are independent and uncommon, Croston’s method—which was developed for predicting intermittent demand—is unsuitable for continuous data, such as inflation, leading to skewed and unstable forecasts. Because linear regression assumes continuous relationships between variables, it ignores the nonlinearities and lag feedback effects that characterize real-world inflation.

Underfitting results from this rigidity, especially when the economy changes as a result of external shocks, policy adjustments, or shifts in commodity prices. These flaws demonstrate how static or oversimplified models fail to adequately represent the complex, time-varying nature of inflation. Inflation in the US is influenced by structural shifts, external disturbances, and shifting policy responses, all of which require dynamic model learning and adaptation. Because of their data-driven adaptability, ANN and hybrid techniques thus exhibit clear advantages over traditional frameworks in controlling such complicated dynamics.

There are important policy ramifications to the increased predictive capacity of ANN-based models. In order to forecast shifts in the structural regime and inflationary pressures, central banks can use these models as early warning systems. By exposing nonlinear links, they facilitate faster and more fact-based monetary policy adjustments. It is challenging to convey policies in a way that is understandable due to the restricted interpretability of neural networks, notwithstanding their accuracy. To get accuracy and transparency, it is feasible to combine ANN forecasts with interpretable statistical models, like ARIMA or ETS.

The necessity of finding a balance between generalizability and flexibility is highlighted by the overfitting observed in complex models like TBATS and Bagged Ensembles. Even while these models are capable of capturing complex seasonal and structural characteristics, if they are too complicated and replicate noise rather than actual trends, their usefulness for future forecasting may be reduced. For accurate inflation modeling, frameworks that balance responsiveness and stability must be adaptable yet cost-effective.

Overall, the results demonstrate that the naturally nonlinear inflation dynamics in the US are influenced by a number of interacting influences. Adaptive and hybrid models that account for this complexity, particularly ANN-based methods, yield the most accurate predictions. However, statistical or transformation-based models alone are insufficiently responsive to capture shifts in the structure and behavior of the economy. The aforementioned findings highlight how important it is to combine flexible, data-driven methodologies with easily understood frameworks in order to increase the precision and policy relevance of inflation forecasts.

The forecast for inflation of the next 10 months of inflation of the USA is shown in Fig. 5.

ANN model forecast of U.S. inflation from June 2024 to March 2025, showing predicted values with 80% and 95% confidence intervals.
Fig. 5.
ANN model forecast of U.S. inflation from June 2024 to March 2025, showing predicted values with 80% and 95% confidence intervals.

The smooth fitted line of the ANN model closely resembles the real series in Fig. 5, demonstrating how well it captures the underlying inflation dynamics. The projection for June 2024–March 2025, based on the trained model, indicates a slow increase in inflation. The forecasted inflation rate, which is also displayed in Table 3, increases from roughly 2.32% in June 2024 to 2.37% in March 2025, suggesting modest inflationary pressure in the upcoming months.

Table 3. Forecast the 10-month inflation of the USA from June 2024 to March 2025.
Months Forecasts
June(2024) 2.3163
July 2.3197
Aug 2.3282
Sep 2.3376
Oct 2.3447
Nov 2.3523
Dec 2.3585
Jan(2025) 2.3642
Feb 2.3688
March 2.3717

Both the 80% and 95% prediction ranges are shown in the forecast graphic as shaded bands surrounding the forecast line. The degree of prediction uncertainty is reflected in these intervals. Toward the start of the forecast horizon (June–August 2024), the bands are comparatively narrow, suggesting strong model confidence in short-term forecasts. But as the time horizon grows toward early 2025, the gaps get wider, showing that forecast uncertainty rises with time. This is a typical occurrence in time series modeling because of compounding unpredictability and data noise.

The predicted time is visually separated from the historical data by the darkened grey area in the plot. Within this range, the projection is steady and does not alter suddenly, suggesting that the ANN model anticipates modest monthly rises in inflation.

All things considered, the ANN model produces accurate short-term predictions and realistic medium-term estimates. However, the increasing prediction intervals emphasize how crucial it is to exercise caution when evaluating longer-term projections because these expected values may deviate due to unforeseen structural changes, policy changes, or economic shocks.

6.1 Limitations of the short-term forecast horizon

The ANN model yields accurate short-term forecasts, but over longer time horizons, its accuracy declines. Although the short-term estimate from June 2024 to March 2025 looks solid, it might not adequately account for sudden policy changes, global shocks, or overfitting problems, underscoring the need for frequent updates and additional economic study.

7. Modelling Inflation Rate Data using Probability Models

Probability distributions play a crucial role in capturing variability, uncertainty, and the probabilistic nature of temporal processes in time series data, thereby supporting robust analysis, forecasting, and decision-making. In recent years, traditional lifetime distributions such as the Weibull and inverse Weibull have been among the most widely applied. These distributions typically exhibit left-skewed density behavior, consistent with data histograms that also show left skewness. In such cases, the left (negative) tail of the distribution is longer than the right, and the mean tends to be lower than the median, suggesting that negatively skewed distributions provide a suitable fit. To model these data, a range of distributions can be employed, including the W, EW, KW, E, and IW distributions. The flexibility of their hazard rate (HR) functions—ranging from monotone to non-monotone—has further driven their broad applicability across diverse fields.

Table 4 provides the maximum likelihood estimates (MLEs) of the model parameters with their standard errors (Ses), along with the corresponding information criteria values Akaike information criterion (AIC), consistent Akaike information criterion (CAIC), Bayesian information criterion (BIC), and Hannan-Quinn information criterion (HQIC) for the competing distributions (W, EW, KW, E, and IW). From these results, it is evident that the W distribution outperforms its counterparts, as it records the lowest values across all information criteria. In addition, its parameter estimates show relatively small standard errors, indicating reliable estimation. The density plots further support these findings, showing that the W distribution offers the best fit to the observed data. Therefore, the Weibull distribution emerges as the most appropriate model for capturing the dynamics of U.S. seasonal inflation data.

Table 4. Information criteria and MLEs with SEs for W, EW, KW, E, and IW distributions.
Models AIC CAIC BIC HQIC MLEs
W 233.9582 234.0054 241.0563 236.8127

0.0046(0.00075)

6.6877(0.1945)

EW 234.2634 234.3583 244.9107 238.5452

0.0044(0.00086)

6.67816(0.2100)

0.9173(0.0766)

KW 245.7818 245.9406 259.9782 251.4909

0.0084(0.0026)

4.4109(0.7727)

1.5896(0.3032)

8.5516(9.4220)

IW 692.3666 692.4139 699.4648 695.2212

2.9889(0.1906)

1.9794(0.0648)

E 894.3087 894.3244 897.8578 895.7359 0.4790(0.0299)

Fig. 6 displays the fitted W, EW, KW, E, and IW distributions. While each distribution shows slight differences in tail behavior, peak height, and width, the W distribution provides the closest fit to the data and demonstrates superior performance compared to the others.

Estimated density plots of USA seasonal inflation data.
Fig. 6.
Estimated density plots of USA seasonal inflation data.

8. Discussion

This study examines and contrasts a variety of methods for predicting inflation in the United States, including probabilistic methods, stochastic linear and nonlinear time series models, and hybrid frameworks. Using monthly data from January 2003 to September 2024, the analysis evaluates how well different models—including ANNs—predict outcomes. Standard statistical criteria, such as MAE, MAPE, MASE, and RMSE, were used to assess model accuracy in order to guarantee a thorough comparison of forecasting capacity. The ANN showed the lowest error values (RMSE = 0.1132, MAE = 0.0809, MAPE = 4.7993, MASE = 0.2298) among all competing approaches, demonstrating its excellent ability to minimize forecast errors and accurately represent the nonlinear and changing nature of inflation dynamics.

The results confirm that ANNs perform better than traditional and hybrid methods, making them strong and trustworthy instruments for forecasting changes in inflation in intricate economic contexts.

Along with forecasting, the study uses probabilistic modeling to investigate the seasonal nature of U.S. inflation. Several probability distributions, such as the W, EW, KW, E, and IW distributions, were used to reflect the data’s inherent variability and cyclical character. These models’ appropriateness was evaluated using information-based criteria such as HQIC, AIC, CAIC, and BIC. The Weibull model, which fit the seasonal inflation data the best overall, had the lowest values across all criteria among the investigated distributions. The EW, KW, E, and IW distributions, on the other hand, yielded relatively larger values, indicating a weaker fit to the patterns in the data. These results show that the Weibull distribution is the best suitable option for capturing the probabilistic nature of U.S. inflation dynamics because of its exceptional adaptability to asymmetric and heavy-tailed features of inflation.

According to the overall evaluation based on forecasting accuracy metrics (RMSE, MAE, MAPE, and MASE) and information criteria (AIC, CAIC, BIC, and HQIC), the models with the lowest values across these benchmarks offer the most accurate and reliable depiction of inflation trends. These models provide important insights for economic forecasting and policymaking by improving predicted accuracy and advancing our grasp of the structural characteristics underpinning inflation.

A more thorough knowledge of inflation behavior can be obtained by combining probabilistic modeling with time series forecasting, especially when it comes to reflecting the risk and uncertainty of future price changes. Even though the main goal of time series models like ARIMA, ETS, and ANN is to find trends, seasonality, and temporal dependencies, they frequently provide deterministic predictions that might not accurately represent the spectrum of potential future events. Conversely, probabilistic models use the statistical distribution of past data to estimate the likelihood of different inflation scenarios, such as moderate inflation, disinflation, or abrupt price spikes, in order to quantify uncertainty.

This combination improves inflation analysis’s interpretive and predictive capabilities from an economic standpoint. While probabilistic models explain the distributional characteristics and tail behaviors that indicate volatility or the probability of extreme outcomes, time series projections offer point estimates that aid in tracking inflationary trends over time. For example, researchers can detect asymmetries in inflation dynamics—where positive or negative shocks may have varying magnitudes and persistence—by using the Weibull distribution’s flexibility in simulating skewness and heavy tails. These kinds of insights are essential for policymakers because they show the likelihood of deviations from goal levels in addition to the anticipated inflation trend.

Analysts can create reliable and useful projections for decision-making by integrating these two frameworks. By converting statistical results into risk-based metrics, the probabilistic viewpoint aids in the creation of monetary policies that take uncertainty into consideration. In order to inform interest rate adjustments and inflation-targeting tactics, central banks, for instance, can use the probability density functions to determine the possibility that inflation will surpass particular thresholds. Thus, by connecting statistical variability to practical policy implications, risk assessment, and strategic planning, the combination of probabilistic and time series modeling enhances not just technical performance but also the economic interpretation.

9. Conclusions

The primary aim of this study was to enhance inflation forecasting accuracy by combining advanced hybrid and non-hybrid time-series models with rigorous statistical methods. Compared with ARIMA (RMSE = 0.1268, MAPE = 5.38%) and ETS (RMSE = 0.1339, MAPE = 5.51%), the ANN model produced the lowest RMSE (0.1132), MAE (0.0809), MAPE (4.80%), and MASE (0.2298), confirming its superior predictive performance. In addition, model selection criteria—including AIC (233.9582), CAIC (234.0054), BIC (241.0563), and HQIC (236.8127)—indicated that the Weibull distribution was the most suitable probabilistic model for capturing seasonal inflation patterns and modeling the observed fluctuations.

The ANN-based forecasts reveal a steadily rising inflation trajectory with clear seasonal peaks, providing economists, financial analysts, and policymakers with actionable, data-driven insights for anticipating inflationary pressures and designing effective policy responses. By integrating modern computational approaches with traditional statistical techniques, this study establishes a robust forecasting framework that enhances methodological development while offering practical tools for economic decision-making.

Future research could extend the hybrid forecasting framework by incorporating deep-learning architectures such as LSTM (long short-term memory) or GRU (gated recurrent unit) to capture more complex nonlinear dynamics. Additional studies may also evaluate model performance using post-2024 data, compare forecasting accuracy across countries or historical periods, or integrate macroeconomic covariates to further assess model generalizability. Overall, this study makes a substantial contribution by advancing state-of-the-art forecasting methodology and delivering a practical, high-accuracy framework that supports real-world inflation monitoring and policy formulation.

CRediT authorship contribution statement

Muzna Sarwar: Conceptualization, methodology, software, data interpretation, writing original draft, writing, review & editing; L. S. Diab: Conceptualization, methodology, writing original draft, data interpretation, writing, review & editing; Farrukh Jamal: Conceptualization, software, data interpretation, writing original draft, writing, review & editing; Zawar Hussain: Conceptualization, writing, review & editing, data interpretation, methodology; Ibrahim Elbatal: Methodology, data interpretation, writing, review & editing; Ahmed Z. Afify: Methodology, data interpretation, software, project administration, writing, review & editing. All authors have read and agreed to the published version of the manuscript.

Declaration of competing interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Declaration of generative AI and AI-assisted technologies in the writing process

The authors confirm that there was no use of artificial intelligence (AI)-assisted technology for assisting in the writing or editing of the manuscript, and no images were manipulated using AI.

Funding

This work was supported and funded by the Deanship of Scientific Research at Imam Mohammad Ibn Saud Islamic University (IMSIU) (grant number IMSIU-DDRSP2602).

References

  1. , , . Diffusion index-based inflation forecasts for the euro area. working paper series 2001:61. SSRN J. https://doi.org/10.2139/ssrn.356081
    [Google Scholar]
  2. . Stochastic models for forecasting inflation rate: Empirical evidence from Romania. Procedia Econ Financ. 2015;20:44-52. https://doi.org10.1016/S2212-5671(15)00045-3
    [Google Scholar]
  3. , , , , , . Determinants of inflation in Pakistan: an econometric analysis using Johansen co-integration approach. Aust J Bus Manag Res. 2014;4:7182. https://doi.org/10.52283/NSWRCA.AJBMR.20110105A09.
    [Google Scholar]
  4. Camba-Méndez, G., Kapetanios, G., 2004. Forecasting euro area inflation using dynamic factor measures of underlying inflation. Working Paper Series 402, European Central Bank.
  5. , . Empirical study of seasonal unit roots in forecasting. Int J Forecast. 1997;13:341-355. https://doi.org/10.1016/S0169-2070(97)00022-8
    [Google Scholar]
  6. . A positive theory of inflation and inflation variance. Econ Inquiry. 1989;27:105-116. https://doi.org/10.1111/j.1465-7295.1989.tb01166.x
    [Google Scholar]
  7. . Seasonality, non-stationarity, and the forecasting of monthly time series. Int J Forecast. 1991;7:199-208. https://doi.org/10.1016/0169-2070(91)90054-y
    [Google Scholar]
  8. . Exponential smoothing: The state of the art. Journal of forecasting. 1985;4:1-28. https://doi.org/10.1002/for.3980040103
    [Google Scholar]
  9. Hofmann, B., 2008. Do monetary indicators lead euro area inflation? Working Paper Series, 867, European Central Bank. SSRN J. https://doi.org/10.2139/ssrn.1090542
  10. . Forecasting euro area inflation: Does aggregating forecasts by HICP component improve forecast accuracy? Int J Forecast. 2005;21:119-136. https://doi.org/10.1016/j.ijforecast.2004.04.005
    [Google Scholar]
  11. . Seasonality in regression. Orlando: Academic Press; .
  12. , . Determinants of inflation: A case of Pakistan. J Econ. 2010;1:45-51. https://doi.org/10.1080/09765239.2010.11884923
    [Google Scholar]
  13. . Unit‐root versus deterministic representations of seasonality for forecasting. In: , , eds. A Companion to Economic Forecasting. Chichester: John Wiley & Sons; . https://doi.org/10.1002/9780470996430.ch18
    [Google Scholar]
  14. , , . Seasonal unit roots and forecasts of two-digit European industrial production. Int J Forecast. 1999;15:27-47. https://doi.org/10.1016/s0169-2070(98)00055-7
    [Google Scholar]
  15. , , . Mean shifts, unit roots, and forecasting seasonal time series. Int J Forecast. 1997;13:93-130. https://doi.org/10.1142/9789812778963_0003
    [Google Scholar]
  16. , . Forecasting inflation. J Monetary Econ. 1999;44:293-335. https://doi.org/10.1016/s0304-3932(99)00027-6
    [Google Scholar]
  17. Suseno, S.A., 2009. Inflasi (Seri Kebanksentralan No. 22). Jakarta: Pusat Pendidikan dan Studi Kebanksentralan, Bank Indonesia. Retrieved from https://www.bi.go.id/id/edukasi/Documents/22.%20Inflasi.pdf.
Show Sections