Next Article in Journal
Decoding Pension Funds: Sustainability Indicators for Annual Report Analysis
Previous Article in Journal
Strategic Budgeting and Budgeting Evaluation Effects on China’s Manufacturing Companies’ Performance
Previous Article in Special Issue
Economic News, Social Media Sentiments, and Stock Returns: Which Is a Bigger Driver?
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Professional Forecasters vs. Shallow Neural Network Ensembles: Assessing Inflation Prediction Accuracy

by
Jane M. Binner
1,*,†,
Logan J. Kelly
2,† and
Jonathan A. Tepper
3,4,†
1
Birmingham Business School, University of Birmingham, Birmingham B15 2TY, UK
2
School of Business and Economics, University of Wisconsin, River Falls, WI 54022, USA
3
Perceptronix Limited, Derby DE65 5AE, UK
4
Aston Business School, Aston University, Birmingham B4 7UP, UK
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
J. Risk Financial Manag. 2025, 18(4), 173; https://doi.org/10.3390/jrfm18040173
Submission received: 17 February 2025 / Revised: 15 March 2025 / Accepted: 17 March 2025 / Published: 25 March 2025
(This article belongs to the Special Issue Forecasting and Time Series Analysis)

Abstract

:
Accurate inflation forecasting is crucial for effective monetary policy, particularly during turning points that demand policy realignment. This study examines the efficacy of dedicating ensembles of shallow recurrent neural network models to different forecasting horizons for predicting U.S. inflation turning points more precisely than traditional methods, including the Survey of Professional Forecasters (SPF). We employ monthly data from January 1970 to May 2024, training these ensemble models on information through December 2022 and testing on out-of-sample observations from January 2023 to May 2024. The models generate forecasts at horizons of up to 16 months (one ensemble per horizon), accounting for both short- and medium-term dynamics. The results indicate that such ensembles of recurrent neural networks consistently outperform conventional approaches using key performance metrics, notably detecting inflation turning points earlier and projecting a return to target levels by May 2024—several months ahead of the Survey of Professional Forecasters’ average forecast. These findings underscore the value of such ensembles in capturing complex nonlinear relationships within macroeconomic data, offering a more robust alternative to standard econometric methods. By delivering timely and accurate forecasts, dedicated ensembles of shallow recurrent neural networks hold great promise for informing proactive policy measures and guiding decisions under uncertain economic conditions.

1. Introduction

Accurate inflation forecasting is pivotal for effective monetary policymaking, particularly when a shift in policy stance is warranted to determine turning points in the economy. In this study, we compare the performance of Shallow Neural Network Ensemble Models with the Federal Reserve Bank of Philadelphia’s Survey of Professional Forecasters (SPF) average forecast, focusing on whether or not these ensemble methods provide enhanced predictive accuracy for U.S. inflation. Our findings indicate that these ensemble models are significantly more accurate than the SPF and also identify a break in trend inflation several months ahead of the SPF. Demonstrating improvements relative to the SPF’s widely followed benchmark thus underscores a meaningful contribution to current forecasting practices.
Historically, a robust literature base supports the importance of better forecasting in economics and management science, especially under conditions of financial volatility (Dia, 2001; Binner et al., 2004a, 2004b; Tsay, 2005; G. P. Zhang & Qi, 2005; Bodyanskiy & Popov, 2006; Curry, 2007; Cao et al., 2012; Binner et al., 2024). A long-held macroeconomic hypothesis posits that a stable relationship exists between the growth rate of the money supply and the growth rate of prices (Friedman & Schwartz, 1970; Barnett, 1997). Conventional econometric methods have scrutinized this linkage extensively, yet evidence of nonlinear effects and structural breaks (Bachmeier et al., 2007) has led to an exploration of artificial neural networks (ANN) as an alternative (Virili & Freisleben, 2000; Giles et al., 2001). Such methods potentially avoid pitfalls associated with cointegration tests in the presence of evolving payment technologies or nonstationary conditions (McCallum, 1993).
Traditional statistical models of time-series forecasting enjoyed considerable success from the 1950s onwards, beginning with exponential smoothing (Brown, 1956; Holt, 1957) and progressing to the Auto-Regressive Moving Average (ARMA) model (Whittle, 1951). Box and Jenkins (1970) extended ARMA to ARIMA(p,d,q), addressing nonstationarity by incorporating a differencing parameter, (d), and later introduced seasonal adjustments (SARIMA) to account for seasonality. These innovations brought systematicity and reproducibility to time-series forecasting, and many practitioners subsequently migrated from univariate approaches to multivariate variants such as ARIMAX and SARIMAX, which incorporate exogenous variables. Further advancements allowed multiple input-output relationships in Vector AutoRegressive (VAR), Vector Moving Average (VMA), and Vector Autoregressive Moving Average (VARMA) models (Sims, 1980). Indeed, ARIMA and VAR models have demonstrated success in macroeconomic forecasting (Stock & Watson, 2007; Dede Ruslan et al., 2018). However, a thorough comparison of these linear techniques with complex machine learning methods by Makridakis et al. (2018) and Makridakis and Hibon (2000) revealed that simple linear models often outperformed nonlinear models in small-sample settings. Later work by Cerqueira et al. (2019) showed that machine learning methods may indeed surpass linear benchmarks as sample sizes grow, potentially overcoming the tendency to overfit. More recently, Orojo et al. (2023) found that a multi-recurrent neural network (MRN) could outperform traditional methods when dealing with more complex time-series data.
Despite their effectiveness, linear models such as ARIMAX, SARIMAX, and VAR face several challenges, including assumptions of linearity, stationarity, and stable structural relationships (Petrica et al., 2016). Nonlinear methods, by contrast, can capture dynamic patterns arising from disruptive technological changes or macroeconomic shocks (Kelly et al., 2024; Narayanaa et al., 2023; Nakamura, 2005; Chen, 2020). Markov switching or regime switching models address some nonlinearities by allowing parameters to vary across regimes (Goldfeld & Quandt, 1973; Hamilton, 1988, 1989, 1994), but they rely on predefined transition probabilities and stationary Markov chains, which can limit flexibility (Briggs & Sculpher, 1998). Recurrent neural networks (RNNs), including Long Short-Term Memory (LSTM) and Gated Recurrent Units (GRU), may sidestep some of these constraints by offering continuous, dynamic memory across regimes (Hochreiter & Schmidhuber, 1997; Lin et al., 1996; Tepper et al., 2016). However, RNNs can be computationally intensive (Bengio et al., 1994) and have historically struggled with vanishing gradients, prompting research into architectures featuring memory cells and gating mechanisms (Cho et al., 2014).
In recent years, attention has shifted towards more complex network architectures such as Transformers (Vaswani et al., 2017), which leverage attention mechanisms, and excel at handling large-scale data in parallel (Roumeliotis & Tselikas, 2023). Surprisingly, several studies suggest that Transformers may not consistently outperform simpler RNN variants in time-series forecasting (Murray et al., 2022; Zeng et al., 2022). Moreover, shallow feedforward neural networks remain appealing due to their relative simplicity and generally robust function approximation capabilities (Rumelhart et al., 1986; Goodfellow et al., 2016).
Against this background, we investigate whether an enhanced RNN—a member of the simple recurrent network class that incorporates a temporally sensitive memory mechanism, can more accurately forecast inflation turning points than conventional approaches, including the SPF’s average forecast. We use U.S. data spanning January 1970 to May 2024, training and validating our model from January 1970 to December 2022, and then generating out-of-sample forecasts for January 2023 to May 2024. We dedicate different RNN models to producing a particular step-ahead forecasts from t + 1 to t + 16, enabling an assessment of both short-term and medium-term inflation dynamics. By examining whether critical predictive signals embedded in historical data can improve turning-point detection, our research aims to offer new insights into the utility of machine learning—specifically ensembles of shallow RNNs dedicated to each forecast horizon, for guiding monetary policy decisions. Our principal conclusion is that these ensemble methods provide a robust alternative to more traditional econometric and survey-based forecasts, thereby contributing to the ongoing debate on how best to capture the complexities of modern macroeconomic data.
A substantial and expanding body of research investigates the mechanisms behind inflation expectations formation, particularly through the lens of rational inattention. Rational inattention refers to the infrequent updating of information sets by firms, households, and professional forecasters, which affects their ability to accurately and promptly predict inflationary changes (Coibion et al., 2018). This concept highlights why expectations adjust gradually rather than instantaneously to new economic information, posing significant challenges for policymakers who rely on timely forecasts to guide economic policy decisions.
The inherent sluggishness of expectation formation processes, particularly among policymakers, significantly complicates the timely implementation of economic policies. To mitigate these challenges, recent studies suggest that advanced forecasting techniques, such as Shallow Neural Network Ensembles, may offer improved predictive capabilities by rapidly integrating new information. For example, D. Zhang et al. (2010) demonstrated the effectiveness of neural network models in forecasting turning points within China’s business cycles, underscoring the potential utility of such methods for predicting critical economic shifts.

2. Materials and Methods

2.1. Data

We begin with an overview of the data. Figure 1 shows the month over month percentage change and level of the U.S. consumer price index (CPI), the month over month percentage change in U.S. monetary aggregate, Divisia M4 (DM4) and Divisia M2 (DM2), as well as the 3-month T-Bill Rate. Divisia M4 is a broad money measure published by the Center for Financial Stability, CFS, New York (see Belongia & Binner, 2000, for details on the construction and interpretation of evidence on the macroeconomic performance of Divisia monetary aggregates for eleven countries). Divisia M2 use applies the same Divsia monetary aggregation method to monetary assets include in the US Federal Reserve’s M2 definition.

2.2. The Multi-Recurrent Network Methodology

Machine learning techniques are concerned with general pattern recognition or the construction of universal function approximators, Cybenko (1989), that can detect relationships in the data in situations where no obvious patterns exist. The power of neural network approximations is summarized in the well-known theorem of Hornik et al. (1989), which essentially states that a neural network can approximate all probability density functions supported compactly on the real numbers. In view of the Dorsey (2000) contention that standard econometric methods have provided no concrete evidence one way or the other to explain the link between money growth and inflation, we believe neural networks offer more promise in the context of econometric modeling than standard linear models. In essence, this approach helps to define the parameter space in a framework akin to a regime-switching model where the regimes themselves are nonlinear relationships. Although this is somewhat unusual in this context, the application of neural networks in economics and finance is growing, as indicated by the wide range of studies surveyed in Binner et al. (2004a, 2004b, 2005, 2010, 2018, 2024).
A key limitation of the traditional neural network models used for inflation forecasting by Gazely and Binner (2000), Dorsey (2000), and Binner et al. (2005) is that they rely on the feed-forward multilayered perceptron (FF-MLP) trained with standard backpropagation (Rumelhart et al., 1986). Although the limitations of the backpropagation algorithm are well-known, it is the feed-forward nature of the architecture that restricts the types and robustness of temporal dependencies these models can learn and represent. Typically, historical input information must be provided as a temporal input window, which is pre-determined (e.g., a six-month window requiring six months of past values for each explanatory variable). Determining the optimum window size must be carried out empirically, and the strictly forward flow of information confines these models to acting as nonlinear auto-regressive (NLAR) models. Should a moving average component or previous output behavior be essential for the learning task, more complex architectures are required that introduce feedback connections.
Building on our previous successes (Kelly et al., 2024; Tepper et al., 2016; Orojo et al., 2019, 2023), we use a state–of-the-art multi-recurrent network (MRN) in this paper. The backpropagation learning rule (Rumelhart et al., 1986) and its variants are used to train these neural networks (including the MRN) and is considered a powerful optimization technique enabling sophisticated curve fitting (Refenes & Azema-Barac, 1994). A main tenet of backpropagation is its aim to minimize the difference between the network’s predicted output and the actual target output, akin to finding a function that best fits a set of data points.
The curve-fitting ability of an MRN is directly influenced by hyperparameters such as the number of hidden layers and the number of neurons in each layer, analogous to choosing the order of a polynomial. An insufficiently complex model leads to poor fit and prediction, whereas an overly complex model risks overfitting and weak generalization. Neural networks are inductive, meaning they can learn from empirical regularities even when the precise governing rules of a phenomenon are not known. This property is advantageous in economic contexts, where the goal is often to model highly complex systems while retaining predictive power. The learning process allows the network, in effect, to disregard excess input variables. Hence, our methodology requires a large amount of rigor in terms of the machine learning model fitting and selection process.
Our methodology aligns with the curve-fitting perspective outlined by Refenes and Azema-Barac (1994) rather than with traditional statistical approaches, such as those used by Cont (2001) in examining various statistical properties of asset returns. Cont’s (2001) work deals more with distributional properties, tail dynamics, and nonlinear dependence over time. Traditional back testing, commonly applied in statistical models for financial market trading, can obscure overfitting when used with recurrent neural networks (Shirdel et al., 2021). As a result, the orthodox approach to mitigating overfitting in models like the MRN involves regularization (to diminish the impact of unimportant features) and ensembling (combining predictions from multiple models to reduce variance). These advantages will be further considered when discussing the MRN’s architecture and forecasting methodology in the following section. We follow a robust model fitting and selection methodology consistent with Orojo et al. (2023) and also introduce an ensemble-based technique to minimize variance, in conjunction with L2 regularization, as stated above.
This class of RNN extends the traditional FF-MLP architecture to include recurrent connections that allow network activations to feedback as inputs to units within the same or preceding layer(s). Such internal memories enable the RNN to construct dynamic internal representations of temporal order and dependencies which may exist within the data. Units that receive feedback values are often referred to as context or state units. Also, assuming nonlinear activation functions are used, the universal function approximation properties of FF-MLPs naturally extends to RNNs. These properties have led to wide appeal of RNNs for modeling economic time-series data, Ulbricht (1994); Moshiri et al. (1999); Tenti (1996); Tino et al. (2001); Binner et al. (2004a, 2004b; 2010), Cao et al. (2012) and Tepper et al. (2016).
The MRN architecture is shown in Figure 2 and can be expressed generally as:
y ^ t + τ = g f c t , x t ,   W f t , W f c t , W f b t , W g t , W g b t
where y ^ ( t + τ ) denotes the predicted month-on-month percentage change in inflation for a particular forecast horizon, τ, where we explore horizons τ = 1 to 16. As described in Section 2 and later in this Section, the external input vector x ( t ) consists of between two to five input variables representing combinations of: the previous month-on-month percentage change in inflation, natural log of the price level, 3-month treasury bill secondary market rate, and month-on-month growth rates (%) for DM2 or DM4.
c ( t ) , the context vector, forms a complex internal memory state vector that is the concatenation of the: previous hidden state vector (h ( t 1 ) ) ; previous output vector ( y ^ ( t 1 ) ) and the previous external input vector, (x ( t 1 ) ) . We extend each allowable feedback by additional banks of context units (memory banks) on the input layer. The number of additional memory banks relates directly to the degree of granularity at which past and current information is integrated and stored and enables the network to implement a form of sluggish state-based memory to inform decision making over variable-width time delays. Following Ulbricht (1994), Binner et al. (2010), Tepper et al. (2016) and Orojo et al. (2023), we use γ memory banks where γ = 4 as experience has shown that moving beyond this does not lead to enhanced performance for the problems addressed. c ( t ) represents an internal memory of varying rigidity; that is some context units will represent information from very recent time steps and thus change rapidly whilst others will represent information further back in time and change much more slowly. To achieve this, the unit values fed back from the hidden and output layers are combined, to varying degrees, with their respective context units, c i ( t ) , in the input layer, as determined by the weighting values applied to its recurrent links. The influence of the previous context unit values is determined by the weighting values applied to the self-recurrent links. When the weighting on the recurrent links is greater than those on the self-recurrent links then more flexible memories are formed storing more recent information at the expense of historical information stored in the context units. Conversely, if the weighting values on the self-recurrent links are greater than those on the layer recurrent links then more rigid memories are formed since more historical information is preserved in the context units at the expense of the more recent information being fed back from subsequent layers. It is these layer- and self-recurrent link ratios that collectively implement a sluggish state memory whereby network outputs (or decisions) are based on variable integrations of recent and distant past unit activations from the input, hidden and output layers. This weighted integration of current and past information can be generally expressed as follows:
c i t = f ( a j t 1 v j + c i t 1 z i )
where c i ( t ) , refers to context unit i at time t; a j ( t 1 ) denotes either an output activation, hidden unit activation or external input value at time t − 1; v j refers to the connection strength of the recurrent link from a j ( t 1 ) to c i ( t ) where v j = 1 γ j where j = 1,2 , , γ ; c i ( t 1 ) refers to context unit i at time t − 1; and finally z i refers to the connection strength of the self-recurrent link for c i where z i = 1 v j . x ( t ) is the external vector of input variables.
W f t is the weight matrix connecting the external input vector to the hidden layer; W f c t is the weight matrix connecting the context vector to the hidden layer; W f b t is the weight matrix between a single constant input unit, equal to 1, and the hidden layer (representing the hidden threshold or bias term); W g t is the weight matrix connecting the hidden layer to the output layer and W g b t is the weight matrix between a single constant unit, equal to 1, and the output layer, again representing the output threshold or bias term. The vector function, f, and function g return the activation vectors from the hidden and output layers, respectively. We apply the hyperbolic tangent function, σ , to the inner products performed for f to compute the hidden unit values as shown below for a single hidden unit, h j ( t ) :
h j ( t ) = σ i W j i f x i ( t ) + k W j k f c c k ( t ) + W j f b
where σ u = σ e u e u e u + e u .
The number of hidden units is a significant hyper-parameter for this model and is empirical. Finally, we apply the identity function, ρ , for the inner products performed for g to compute the output unit values as shown below for a single output unit, o l ( t ) :
o l ( t ) = ρ j W j l g h j ( t ) + W l g b
where ρ u = u .
We use backpropagation through time (BPTT) (Werbos, 1990; Williams & Peng, 1990) to minimize the following mean-squared-error (MSE) measure:
M S E = 1 M N t = 1 M l = 1 N o l t y ^ l ( t ) 2
where M refers to the number of input patterns and N is the number of output units. BPTT is an efficient gradient-descent learning algorithm for training RNNs where weight changes are determined by a sequence, or time-lag, of input observations rather than a single observation. Input data were scaled to zero mean, and unit variance. To facilitate convergent learning, a linearly decaying learning rate and momentum schedule are used for all training experiments with an initial learning rate of 0.0001 and momentum term of 0.95. This provides an effective time varying learning rate that guarantees convergence of stochastic approximation algorithms and has proven effective for temporal domains, for example, see Lawrence et al. (2000). To minimize the onset of over-fitting, in addition to the ensemble-based methodology discussed later, we apply L2 weight regularization during training with the lambda constant set at 0.001. We generate independent training sequences directly from the time-series using a time window whose lag size is empirically established, i.e., 120 months for the US inflation problem tackled here. The MRN context units are initialized to known values at the beginning of each sequence of data and each element within a sequence is processed sequentially by the network. Although this restricts the MRN to behave as a finite memory model and means that the MRN must first learn to ignore the first context values of each sequence, it allows for efficient randomized learning.
Although such RNNs and variants thereof have enjoyed some success for the time-series problem domains, we tread with cautious optimism. Due to the additional recurrency, such models inherently contain large degrees of freedom (weights) which may require many training cases to constrain. Also, gradient descent-based learning algorithms are notoriously difficult at finding the optimum solution the architecture is theoretically capable of representing: Bengio et al. (1994) and Sharkey et al. (2000). However, the embedded memory architecture within the MRN may compensate for this deficiency, as demonstrated recently by Tepper et al. (2016) and more recently by Orojo et al. (2023). For a comprehensive account of the strengths and limitations of both feedforward and recurrent neural networks and how to optimize them, please see the classic textbook by Goodfellow et al. (2016).
Based on the feature set used, we construct the following three types of MRN:
(a)
simple MRN consisting of three input variables: the month-on-month percentage change in inflation (cpi_percmom, the auto-regressive term), the natural log of the price level (ln_cpi_level) and 3-Month Treasury Bill Secondary Market Rate (dtb3).
(b)
intermediate MRN which includes (a) plus the month-on-month growth rate (%) for the Divisia monetary measure DM4 (dm4_percmom).
(c)
complex MRN which includes (b) plus the month-on-month growth rate (%) for the Divisia monetary measure DM2 (dm2_percmom).
For each of the above feature sets, we follow the same approach used by Binner et al. (2024) and develop a set of sixteen MRN ensembles where each MRN has a monthly input lag structure spanning 10 years and a single hidden layer with 50 nonlinear units attached to single output layer to predict a particular forecast horizon of the month-on-month percentage change in inflation, i.e., t + 1, t + 2,…, or t + 16. Each ensemble consists of five independently trained MRNs fitted to the in-sample period of January 1970–December 2022 and then evaluated on the out-sample period of January 2023 to April 2024; thus, a total of 80 MRNs were generated for each of the three feature sets. As mentioned earlier, a series of independent training sequences of 120 months are generated directly from the time-series and presented in a random fashion to each MRN during the training process, where the context units are initialized to some known value at the beginning of each sequence, and the forecast generated in response to the last observation in the training sequence being stored as the prediction for that particular sequence.
The out-of-sample forecast for each trained MRN is generated in response to being sequentially fed the first out-of-sample input window containing observations from January 2013 until December 2022. The MRNs trained to predict t + 1 are responsible for the January 2023 forecast, the models trained to predict t + 6 are responsible for June 2023 forecast and so on. Each prediction is an average of the respective five MRNs associated with the dedicated forecast horizon. During the out-of-sample period, no further parameter estimates are obviously allowed. As the sequential input window slides over time, only new input observations are passed through the network as they become available. This forecasting methodology is depicted in Figure 3 below. In terms of backtesting, given our previous research exploring different configurations of training, validation and test samples, we have adopted a more efficient approach to the problem of identifying the appropriate level of model complexity by applying regularization and ensemble-based aggregation. We then apply our fitted models to the unseen out-of-sample data.
This is analogous to an SPF making long term forecasts at the end of December 2022 based only on rules they have derived from studying observations between January 1970 and December 2022.

2.3. Survey of Professional Forecasters (SPF)

The purpose of this study is to contrast the performance of macroeconomic forecasts using artificial neural networks to those made by the methods using human judgment that we represent by the Survey of Professional Forecasters (SPF). The SPF is the oldest quarterly survey of macroeconomic forecasts in the United States. The survey began in 1968 and was conducted by the American Statistical Association and the National Bureau of Economic Research. The Federal Reserve Bank of Philadelphia took over the survey in 1990. The Survey of Professional Forecasters’ is a statistical survey, and the official web page offers the actual releases, documentation, mean and median forecasts of all the respondents as well as the individual responses from each economist; the latter are kept confidential with the use of identification numbers, please see https://www.philadelphiafed.org/surveys-and-data/real-time-data-research/survey-of-professional-forecasters, 15 March 2025. It compiles predictions from a diverse panel of professional economists on key economic indicators, including inflation, real GDP, and unemployment. Specifically, the SPF’s inflation forecasts draw widespread attention from policymakers and market participants, serving as a central benchmark for comparing different predictive methods. Because it aggregates forecasts from experts in academia, industry, and financial institutions, the SPF provides a consensus view of future inflation trends over various time horizons. Data from the SPF are publicly available and are regularly updated, allowing researchers and practitioners to assess forecast accuracy, detect shifts in economic sentiment, and evaluate new forecasting models against a longstanding, reputable standard.
The Survey of Professional Forecasters (SPF) generates 16 separate forecasts corresponding to time horizons from t + 1 through t + 16 for each period. While it might seem beneficial to include SPF data as an additional input in an RNN to enhance ensemble prediction accuracy, doing so is not straightforward. This is because the multi-horizon nature of SPF forecasts requires selecting one specific forecast horizon for input. Unfortunately, there is no clear rationale or empirical evidence favoring the choice of one horizon over another, which complicates the integration of SPF data into the model.

2.4. Forecast Evaluation Procedure

Each of the neural network models is compared to the Survey of Professional Forecasters’ (SPF) average inflation forecast over the simulated out-of-sample period from January 2023 through December 2023, providing robust evidence for or against meaningful improvements in forecasting accuracy. To facilitate this comparison, we employ three primary forecast evaluation metrics:
  • Root Mean Squared Error (RMSE), which measures average forecast errors and emphasizes large deviations.
  • Symmetric Mean Absolute Percentage Error (sMAPE) which expresses errors as a percentage, simplifying comparisons across different scales.
  • Theil’s U Statistic, which compares model performance to a naïve forecast, with lower values indicating better accuracy.
  • Improvement Over Random Walk, which quantifies how much better a forecasting model performs compared to a random walk benchmark, expressed as a proportion

3. Results

3.1. Forecast Evaluation

Table 1 compares five forecasting methods, i.e., simple MRN, intermediate MRN, complex MRN, average MRN, and SPF, across these evaluation metrics. The average MRN evaluation statistics, highlighted in bold, demonstrates the preferred model as it has the lowest RMSE and sMAPE signifying the lowest average forecast errors. It also achieves the lowest Theil’s U (0.104), which suggests that it closely aligns with actual values and performs significantly better than a random walk. The high Improvement score of 0.896 for average MRN reflects an almost 90% reduction in forecast error over the baseline, indicating a substantial predictive advantage.
The simple MRN model also performs well, with an Improvement score of 0.876 and a Theil’s U of 0.124, suggesting it is nearly as accurate as average MRN. Meanwhile, the intermediate MRN and complex MRN exhibit slightly higher error metrics and Theil’s U values, indicating that these models are somewhat less accurate. The SPF model, while showing moderate performance, has a relatively higher Theil’s U of 0.152, suggesting its forecast accuracy is less favorable compared to the average MRN and simple MRN. Overall, the average MRN is the best-performing model in terms of accuracy and predictive value, with the simple MRN following closely behind. These models demonstrate a significant improvement over the random walk, making them reliable choices for forecasting applications.

3.2. Comparison of CPI and Forecasts over Time

Figure 4 illustrates each forecasting model’s projection against the observed inflation rate over the test period. Notably, the MRN-based forecasts all suggest that inflation will return to near target levels by May 2024, in contrast to the Survey of Professional Forecasters (SPF), which predicts a near target-level return in October 2024. This discrepancy holds significant policy implications, as the Federal Reserve has faced criticism for delaying cuts to key policy rates.
To assess any potential bias of the RNN forecast, we estimated the model
F E t = α + ϵ t ,
using ordinary least squares on a sample spanning January 2023 to May 2024 (n = 17). The intercept, which represents the average forecast error, was estimated to be −0.000266 (0.000548). This resulted in a t statistic of t(16) = −0.49, and a p value of 0.63. These findings indicate that the intercept term is not significantly different from zero, suggesting that there is no systematic bias in the RNN forecast estimates.

4. Discussion

The comparative results presented here suggest that ensembles of shallow MRNs dedicated to different forecast horizons, in some analyses, can offer notable improvements over conventional forecasts, such as those provided by the Survey of Professional Forecasters (SPF). Specifically, these models consistently project a return to near-target inflation several months earlier than the SPF, underscoring the potential value of machine learning approaches for capturing dynamic, nonlinear relationships in macroeconomic data. These findings align with previous studies that highlight the limitations of linear econometric models, e.g., ARIMA, VAR, when confronted with structural shifts or unanticipated shocks, please see Cerqueira et al. (2019) and Orojo et al. (2023) and more recently, Kelly et al. (2025) using UK inflation data. By incorporating advanced memory mechanisms and ensemble averaging, the neural network models appear better equipped to interpret the historical trajectory of inflation, including subtle turning-point signals that might otherwise be missed by simpler or purely judgmental methods.
From a policy perspective, the difference in forecasting timelines could significantly shape decisions regarding interest rates and monetary policy interventions. As reflected in the discussion by Kristin Forbes (2023), adopting a faster or more aggressive tightening strategy may yield less labor market pain and a swifter re-anchoring of inflation expectations, whereas a more gradual approach reduces the risk of abrupt policy overcorrection. Evidence presented here that neural network models can detect turning points earlier than simpler models suggests that real-time deployment of machine learning forecasts may provide central banks with a more proactive warning system for emerging inflationary pressures or impending economic slowdowns.
In the broader context, these results reinforce growing evidence that data-driven, nonlinear modeling approaches can capture complex economic dynamics that traditional frameworks occasionally overlook. Future research might explore how these models fare in international settings or with higher-frequency data, such as weekly or daily price indices. Moreover, extensions could incorporate additional macro-financial variables—like commodity prices, credit conditions, or labor market indicators—to assess how enriched data inputs influence predictive accuracy. Finally, while our study focuses on inflation forecasts, similar ensemble neural network techniques could usefully be applied to other critical macroeconomic indicators, thus expanding their utility in both research and policymaking domains.

5. Conclusions

This study underscores the promise of MRN ensembles as a powerful tool for inflation forecasting, especially during pivotal turning points that demand policy intervention. By fitting ensembles of MRNs on historical data from January 1970 to December 2022 for each forecasting horizon, then testing out-of-sample performance from January 2023 to May 2024, we demonstrate that these models can predict inflation dynamics with greater precision than conventional methods such as the Survey of Professional Forecasters (SPF). Crucially, the average MRN model outperforms both traditional econometric approaches and simpler neural network variants across multiple metrics, including RMSE, sMAPE, and Theil’s U, providing earlier detection of shifts requiring policy adjustments.
Beyond their superior predictive accuracy, MRNs capture nonlinear and time-varying relationships that are often missed by linear frameworks. By leveraging advanced memory mechanisms, these networks offer actionable insights into inflationary trends well before conventional benchmarks, a finding of key relevance to central banks that require reliable forecasts up to 24 months in advance. As Croushore and Stark (2019) note, surpassing the SPF forecasts is considered a significant breakthrough in economic prediction, and the evidence presented here suggests that such MRN ensembles do so consistently.
The importance of these results lies in their ability to guide timely and effective policy responses. Central banks frequently operate under conditions of substantial uncertainty, and the ability to detect turning points earlier can help them introduce measured interventions that reduce the risk of entrenched inflation or destabilizing economic shocks. Given the growing availability of granular, high-frequency data, MRNs’ aptitude for modeling complex dependencies places them at the forefront of a potential paradigm shift—one in which machine learning complements or even supersedes traditional econometric tools.
Future research will further refine the MRN approaches by incorporating additional macroeconomic and financial variables, as well as by exploring ensemble or hybrid methodologies that extend the existing architecture to boost accuracy, robustness and generalizability. For example, experimenting with pruning deliberately over specified MRNs to find the optimum architecture for the problem, as demonstrated by Orojo (2022). Extending these methods to various economic contexts and time horizons would also illuminate the range of scenarios in which MRN ensembles excel. This work has not addressed the welfare costs of inflation, and we have assumed no menu costs as this topic is beyond the scope of this work but could fruitfully be incorporated into a future study, building on, e.g., Hellwig (2008). Ultimately, the insights gleaned from MRNs can foster more responsive, data-driven policy decision making, underscoring their growing importance in the economics and management science toolkit.

Author Contributions

All authors have contributed to the conceptualization, model design and model evaluation criteria. All authors have been equally engaged with the writing and the project administration. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

No new data is created. All data used in this paper are publicly available from the Survey of Professional Forecasters website at this link: https://www.philadelphiafed.org/surveys-and-data/real-time-data-research/survey-of-professional-forecasters, 15 March 2025.

Acknowledgments

The authors wish to thank three anonymous referees and the associate editor for the very helpful comments which have added value and additional insight to this paper.

Conflicts of Interest

Author Jonathan A. Tepper was employed by the company, Perceptronix. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Bachmeier, L., Leelahanon, S., & Li, Q. (2007). Money Growth and Inflation in the United States. Macroeconomic Dynamics, 11, 113–127. [Google Scholar] [CrossRef]
  2. Barnett, W. A. (1997). Which Road Leads to Stable Money? Economic Journal, 107, 1171–1185. [Google Scholar]
  3. Belongia, M. T., & Binner, J. M. (Eds.). (2000). Divisia monetary aggregates: Theory and practice. Palgrave Macmillan. ISBN 978-0333647448. [Google Scholar]
  4. Bengio, Y., Simard, P., & Frasconi, P. (1994). Learning Long-Term Dependencies with Gradient Descent is Difficult. IEEE Transactions on Neural Networks, 5(2), 157–166. [Google Scholar] [CrossRef] [PubMed]
  5. Binner, J. M., Bissoondeeal, R. K., Elger, T., & Mullineux, A. W. (2005). A Comparison of Linear Forecasting Models and Neural Networks: An Application to Euro Inflation and Euro Divisia. Applied Economics, 37, 665–680. [Google Scholar] [CrossRef]
  6. Binner, J. M., Chaudhry, S. M., Kelly, L. J., & Swofford, J. L. (2018). Risky Monetary Aggregates for the USA and UK. Journal of International Money and Finance, 89, 127–138. [Google Scholar]
  7. Binner, J. M., Dixon, H., Jones, B. E., & Tepper, J. A. (2024). A Neural Network Approach to Forecasting Inflation. In UK economic outlook (Spring ed., pp. 8–11). NIESR. Box A. Available online: https://niesr.ac.uk/publications/neural-network-approach-forecasting-inflation?type=uk-economic-outlook-box-analysis (accessed on 15 March 2025).
  8. Binner, J. M., Gazely, A. M., Chen, S. H., & Chie, B. T. (2004a). Financial Innovation and Divisia Money in Taiwan: Comparative Evidence from Neural Network and Vector Error-Correction Forecasting Models. Contemporary Economic Policy, 22(2), 213–224. [Google Scholar]
  9. Binner, J. M., Kendall, G., & Chen, S. H. (2004b). Applications of Artificial Intelligence in Economics and Finance. In T. B. Fomby, & R. C. Hill (Eds.), Advances in econometrics (Vol. 19, pp. 127–144). Emerald Publishing. [Google Scholar]
  10. Binner, J. M., Tino, P., Tepper, J. A., Anderson, R. G., Jones, B. E., & Kendall, G. (2010). Does Money Matter in Inflation Forecasting? Physica A: Statistical Mechanics and its Applications, 389, 4793–4808. [Google Scholar]
  11. Bodyanskiy, Y., & Popov, S. (2006). Neural network approach to forecasting of quasiperiodic financial time series. European Journal of Operational Research, 175, 1357–1366. [Google Scholar]
  12. Box, G. E. P., & Jenkins, G. M. (1970). Time series analysis: Forecasting and control. Holden Day. [Google Scholar]
  13. Briggs, A., & Sculpher, M. (1998). An introduction to Markov modelling for economic evaluation. Pharmaco Economics, 13(4), 397–409. [Google Scholar]
  14. Brown, R. G. (1956). Exponential smoothing for predicting demand. Available online: http://legacy.library.ucsf.edu/tid/dae94e00 (accessed on 13 January 2025).
  15. Cao, B., Ewing, B. T., & Thompson, M. A. (2012). Forecasting Wind Speed with Recurrent Neural Networks. European Journal of Operational Research, 221, 148–154. [Google Scholar]
  16. Cerqueira, V., Torgo, L., & Soares, C. (2019). Machine learning vs. statistical methods for time series forecasting: Size matters. Applied Soft Computing, 80, 1–18. [Google Scholar]
  17. Chen, J. (2020). Economic forecasting with autoregressive methods and neural networks. Available online: https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3521532 (accessed on 15 March 2025).
  18. Cho, K., van Merrienboer, B., Bahdanau, D., Bougares, F., Schwenk, H., & Bengio, Y. (2014). Learning phrase representations using rnn encoder–decoder for statistical machine translation. arXiv, arXiv:1406.1078. [Google Scholar]
  19. Coibion, O., Gorodnichenko, Y., & Kamdar, R. (2018). The formation of expectations, inflation, and the Phillips curve. Journal of Economic Literature, 56(4), 1447–1491. [Google Scholar] [CrossRef]
  20. Cont, R. (2001). Empirical properties of asset returns: Stylized facts and statistical issues. Quantitative Finance, 1(2), 223–236. [Google Scholar] [CrossRef]
  21. Croushore, D., & Stark, T. (2019). Fifty years of the survey of professional forecasters. Available online: https://www.philadelphiafed.org/the-economy/macroeconomics/fifty-years-of-the-survey-of-professional-forecasters (accessed on 15 March 2025).
  22. Curry, B. (2007). Neural Networks and Seasonality: Some Technical Considerations. European Journal of Operational Research, 179, 267–274. [Google Scholar] [CrossRef]
  23. Cybenko, G. (1989). Approximation by superpositions of a sigmoid function. Mathematics of Control, Signals and Systems, 2, 303–314. [Google Scholar] [CrossRef]
  24. Dede Ruslan, D., Rusiadi, R., Novalina, A., & Lubis, A. (2018). Early detection of the financial crisis of developing countries. International Journal of Civil Engineering and Technology, 9(8), 15–26. [Google Scholar]
  25. Dia, H. (2001). An object-oriented neural network approach to traffic forecasting. European Journal of Operational Research, 131, 253–261. [Google Scholar]
  26. Dorsey, R. E. (2000). Neural Networks with Divisia Money: Better Forecasts of Future Inflation? In M. T. Belongia, & J. M. Binner (Eds.), Divisia monetary aggregates: Theory and practice (pp. 12–28). Palgrave Macmillan. [Google Scholar]
  27. Forbes, K. (2023, October 19–20). Monetary policy under uncertainty: The hare or the tortoise? Boston Federal Reserve Bank Conference, Boston, MA, USA. [Google Scholar]
  28. Friedman, M., & Schwartz, A. J. (1970). Monetary statistics of the United States. Columbia University Press. [Google Scholar]
  29. Gazely, A. M., & Binner, J. M. (2000). A neural network approach to the Divisa index debate: Evidence from three countries. Applied Economics, 32, 1607–1615. [Google Scholar]
  30. Giles, C. L., Lawrence, S., & Tsoi, A. C. (2001). Noisy time series prediction using a recurrent neural network and grammatical inference. Machine Learning, 44(1/2), 161–183. [Google Scholar] [CrossRef]
  31. Goldfeld, S., & Quandt, R. E. (1973). A Markov model for switching regressions. Journal of Econometrics, 1(1), 3–15. [Google Scholar] [CrossRef]
  32. Goodfellow, I., Bengio, Y., & Courville, A. (2016). Deep learning. MIT Press. [Google Scholar]
  33. Hamilton, J. D. (1988). Rational expectations econometric analysis of changes in regime: An investigation of the term structure of interest rates. Journal of Economic Dynamics and Control, 12, 385–423. [Google Scholar] [CrossRef]
  34. Hamilton, J. D. (1989). A new approach to the economic analysis of non-stationary time series and the business cycle. Econometrica, 57(2), 357–384. [Google Scholar] [CrossRef]
  35. Hamilton, J. D. (1994). Time series analysis. Princeton University Press. [Google Scholar]
  36. Hellwig, C. (2008). Welfare costs of inflation in a menu cost model. American Economic Review, 98(2), 438–443. [Google Scholar]
  37. Hochreiter, S., & Schmidhuber, J. (1997). Long short-term memory. Neural Computation, 9(8), 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  38. Holt, C. C. (1957). Forecasting seasonal and trends by exponentially weighted moving averages. International Journal of Forecasting, 20, 5–10. [Google Scholar] [CrossRef]
  39. Hornik, K., Stinchcombe, M., & White, H. (1989). Multilayer feedforward networks are universal approximators. Neural Networks, 2(5), 359–366. [Google Scholar] [CrossRef]
  40. Kelly, L. J., Binner, J. M., & Tepper, J. A. (2024). Do Monetary Aggregates Improve Inflation Forecasting in Switzerland? Journal of Management Policy and Practice, 25(1), 124–133. [Google Scholar] [CrossRef]
  41. Kelly, L. J., Tepper, J. A., Binner, J. M., Dixon, H., & Jones, B. E. (2025). National institute UK economic outlook box A: (NIESR): An alternative UK economic forecast. Available online: https://niesr.ac.uk/wp-content/uploads/2025/02/JC870-NIESR-Outlook-Winter-2025-UK-Box-A.pdf?ver=u3KMjmyGBUB2ukqGdvpT (accessed on 19 March 2025).
  42. Lawrence, S., Giles, C. L., & Fong, S. (2000). Natural language grammatical inference with recurrent neural networks. IEEE Transactions on Knowledge and Data Engineering, 12, 126–140. [Google Scholar] [CrossRef]
  43. Lin, T., Horne, B., Tino, P., & Giles, C. L. (1996). Learning long-term dependencies in narx recurrent neural networks. IEEE Transactions on Neural Networks, 7(6), 1329–1338. [Google Scholar]
  44. Makridakis, S., & Hibon, M. (2000). The M3-competition: Results, conclusions and implications. International Journal of Forecasting, 16(4), 451–476. [Google Scholar]
  45. Makridakis, S., Spiliotis, E., & Assimakopoulos, V. (2018). Statistical and machine learning forecasting methods: Concerns and ways forward. PLoS ONE, 13(3), e0194889. [Google Scholar] [CrossRef]
  46. McCallum, B. T. (1993). Unit roots in macroeconomic time series: Some critical issues. Federal Reserve Bank of Richmond Economic Quarterly, 79(2), 13–43. [Google Scholar]
  47. Moshiri, S., Cameron, N., & Scuse, D. (1999). Static, dynamic, and hybrid neural networks in forecasting inflation. Computational Economics, 14, 219–235. [Google Scholar] [CrossRef]
  48. Murray, C., Chaurasia, P., Hollywood, L., & Coyle, D. (2022). A comparative analysis of state-of-the-art time series forecasting algorithms. In International conference on computational science and computational intelligence. IEEE. [Google Scholar]
  49. Nakamura, E. (2005). Inflation forecasting using a neural network. Economics Letters, 86, 373–378. [Google Scholar]
  50. Narayanaa, T., Skandarsini, R., Ida, S. R., Sabapathy, S. R., & Nanthitha, P. (2023, December 21–23). Inflation prediction: A comparative study of arima and lstm models across different temporal resolutions [Conference session]. 2023 3rd International Conference on Innovative Mechanisms for Industry Applications (ICIMIA) (pp. 1390–1395), Bengaluru, India. [Google Scholar] [CrossRef]
  51. Orojo, O. (2022). Optimizing sluggish state-based neural networks for effective time-series processing [Ph.D. Thesis, Nottingham Trent University]. Unpublished. [Google Scholar]
  52. Orojo, O., Tepper, J. A., Mcginnity, T. M., & Mahmud, M. (2019, December 6–9). A multi-recurrent network for crude oil price prediction [Conference session]. 2019 IEEE Symposium Series On Computational Intelligence (SSCI) (pp. 2940–2945), Xiamen, China. [Google Scholar] [CrossRef]
  53. Orojo, O., Tepper, J. A., Mcginnity, T. M., & Mahmud, M. (2023). The multi-recurrent neural network for state-of-the-art time-series processing. Procedia Computer Science, 222, 488–498. [Google Scholar] [CrossRef]
  54. Petrica, A., Stancu, S., & Tindeche, A. (2016). Limitations of arima models in financial and monetary economics. Theoretical and Applied Economics, 23(4), 19–42. [Google Scholar]
  55. Refenes, A. N., & Azema-Barac, M. (1994). Neural network applications in financial asset management. Neural Computing and Applications, 2, 13–39. [Google Scholar]
  56. Roumeliotis, K. I., & Tselikas, N. D. (2023). ChatGPT and Open-AI models: A preliminary review. Future Internet, 15(6), 192. [Google Scholar] [CrossRef]
  57. Rumelhart, D. E., Hinton, G. E., & Williams, R. A. (1986). Learning internal representations by error propagation. In D. E. Rumelhart, & J. L. McClelland (Eds.), Parallel distributed processing: Explorations in the microstructure of cognition (Vol. 1, pp. 318–362). MIT Press. [Google Scholar]
  58. Sharkey, N., Sharkey, A., & Jackson, S. (2000). Are SRNs sufficient for modelling language acquisition? In P. Broeder, & J. M. J. Murre (Eds.), Models of language acquisition: Inductive and deductive approaches (pp. 35–54). Oxford University Press. [Google Scholar]
  59. Shirdel, M., Asadi, R., Do, D., & Hintlian, M. (2021). Deep learning with kernel flow regularization for time series forecasting. arXiv, arXiv:2109.11649. [Google Scholar] [CrossRef]
  60. Sims, C. A. (1980). Macroeconomics and reality. Econometrica, 48(1), 1–48. [Google Scholar]
  61. Stock, J. H., & Watson, M. W. (2007). Why has U.S. inflation become harder to forecast? Journal of Money, Credit and Banking, 39, 3–33. [Google Scholar] [CrossRef]
  62. Tenti, P. (1996). Forecasting foreign exchange rates using recurrent neural networks. Applied Artificial Intelligence, 10, 567–581. [Google Scholar]
  63. Tepper, J. A., Shertil, M. S., & Powell, H. M. (2016). On the importance of sluggish state memory for learning long term dependency. Knowledge-Based Systems, 102, 1–11. [Google Scholar] [CrossRef]
  64. Tino, P., Horne, B. G., & Giles, C. L. (2001). Attractive periodic sets in discrete time recurrent networks (with emphasis on fixed point stability and bifurcations in two–neuron networks). Neural Computation, 13(6), 1379–1414. [Google Scholar] [PubMed]
  65. Tsay, R. S. (2005). Analysis of financial time series. Wiley and Sons. ISBN 9780471690740. [Google Scholar]
  66. Ulbricht, C. (1994, July 31–August 4). Multi-recurrent networks for traffic forecasting [Conference session]. Twelfth National Conference on Artificial Intelligence (pp. 883–888), Seattle, WA, USA. [Google Scholar]
  67. Vaswani, A., Shazeer, N. M., Parmar, N. P., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, L. K., & Polosukhin, I. (2017). Attention is all you need. arXiv, arXiv:1706.03762. [Google Scholar]
  68. Virili, F., & Freisleben, B. (2000). Nonstationarity and data preprocessing for neural network predictions of an economic time series. In Neural networks, 2000. IJCNN 2000, proceedings of the IEEE-INNS-ENNS international joint conference on Como, Italy, July 24–27 (Vol. 5, pp. 129–134). IEEE. [Google Scholar]
  69. Werbos, P. J. (1990). Backpropagation through time: What it does and how to do it. Proceedings of the IEEE, 78(10), 1550–1560. [Google Scholar] [CrossRef]
  70. Whittle, P. (1951). Hypothesis testing in time series analysis. Almquist & Wiksell. [Google Scholar]
  71. Williams, R. J., & Peng, J. (1990). An efficient gradient-based algorithm for on-line training of recurrent network trajectories. Neural Computation, 2(4), 490–501. [Google Scholar]
  72. Zeng, A., Chen, M., Zhang, L., & Xu, Q. (2022). Are transformers effective for time series forecasting? arXiv, arXiv:2205.13504. [Google Scholar]
  73. Zhang, D., Yu, L., Wang, S., & Xie, H. (2010). Neural network methods for forecasting turning points in economic time series: An asymmetric verification to business cycles. Frontiers of Computer Science in China, 4, 254–262. [Google Scholar]
  74. Zhang, G. P., & Qi, M. (2005). Neural network forecasting for seasonal and trend time series. European Journal of Operational Research, 160(2), 501–514. [Google Scholar]
Figure 1. Variables used to estimate the MRN models. Note that the Consumer Price Index (CPI) is seasonally adjusted and based on the 1982–1984 base period, and that the 3-Month T-Bill Rate is expressed as an annual percentage rate.
Figure 1. Variables used to estimate the MRN models. Note that the Consumer Price Index (CPI) is seasonally adjusted and based on the 1982–1984 base period, and that the 3-Month T-Bill Rate is expressed as an annual percentage rate.
Jrfm 18 00173 g001
Figure 2. Multi-recurrent Network Architecture (unit shading indicates different levels of activation for illustrative purposes).
Figure 2. Multi-recurrent Network Architecture (unit shading indicates different levels of activation for illustrative purposes).
Jrfm 18 00173 g002
Figure 3. MRN forecasting methodology.
Figure 3. MRN forecasting methodology.
Jrfm 18 00173 g003
Figure 4. Comparison of CPI inflation and forecasted values over time for each neural network model. CPI is the actual CPI inflation rate. MRN forecast is the forecast of CPI inflation and SPF is the Survey of Professional Forecasters’ forecast of CPI inflation.
Figure 4. Comparison of CPI inflation and forecasted values over time for each neural network model. CPI is the actual CPI inflation rate. MRN forecast is the forecast of CPI inflation and SPF is the Survey of Professional Forecasters’ forecast of CPI inflation.
Jrfm 18 00173 g004
Table 1. Forecast model evaluation statistics.
Table 1. Forecast model evaluation statistics.
Forecast MethodRMSEsMAPETheil U
Statistic
Improvement Over Random Walk
simpleMRN0.47210.91%0.1240.876
intermediateMRN0.63716.47%0.1670.833
complexMRN0.62715.29%0.1640.836
averageMRN0.3959.66%0.1040.896
SPF0.58011.33%0.1520.848
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Binner, J.M.; Kelly, L.J.; Tepper, J.A. Professional Forecasters vs. Shallow Neural Network Ensembles: Assessing Inflation Prediction Accuracy. J. Risk Financial Manag. 2025, 18, 173. https://doi.org/10.3390/jrfm18040173

AMA Style

Binner JM, Kelly LJ, Tepper JA. Professional Forecasters vs. Shallow Neural Network Ensembles: Assessing Inflation Prediction Accuracy. Journal of Risk and Financial Management. 2025; 18(4):173. https://doi.org/10.3390/jrfm18040173

Chicago/Turabian Style

Binner, Jane M., Logan J. Kelly, and Jonathan A. Tepper. 2025. "Professional Forecasters vs. Shallow Neural Network Ensembles: Assessing Inflation Prediction Accuracy" Journal of Risk and Financial Management 18, no. 4: 173. https://doi.org/10.3390/jrfm18040173

APA Style

Binner, J. M., Kelly, L. J., & Tepper, J. A. (2025). Professional Forecasters vs. Shallow Neural Network Ensembles: Assessing Inflation Prediction Accuracy. Journal of Risk and Financial Management, 18(4), 173. https://doi.org/10.3390/jrfm18040173

Article Metrics

Back to TopTop