nep-cmp New Economics Papers
on Computational Economics
Issue of 2025–09–22
35 papers chosen by
Stan Miles, Thompson Rivers University


  1. An Interpretable Deep Learning Model for General Insurance Pricing By Patrick J. Laub; Tu Pho; Bernard Wong
  2. MM-DREX: Multimodal-Driven Dynamic Routing of LLM Experts for Financial Trading By Yang Chen; Yueheng Jiang; Zhaozhao Ma; Yuchen Cao Jacky Keung; Kun Kuang; Leilei Gan; Yiquan Wu; Fei Wu
  3. Automated Trading System for Straddle-Option Based on Deep Q-Learning By Yiran Wan; Xinyu Ying; Shengzhen Xu
  4. DeltaHedge: A Multi-Agent Framework for Portfolio Options Optimization By Feliks Ba\'nka; Jaros{\l}aw A. Chudziak
  5. Identifying economic narratives in large text corpora: An integrated approach using large language models By Schmidt, Tobias; Lange, Kai-Robin; Reccius, Matthias; Müller, Henrik; Roos, Michael W. M.; Jentsch, Carsten
  6. Neural Functionally Generated Portfolios By Michael Monoyios; Olivia Pricilia
  7. A comparative analysis of machine learning algorithms for predicting probabilities of default By Adrian Iulian Cristescu; Matteo Giordano
  8. Finance-Grounded Optimization For Algorithmic Trading By Kasymkhan Khubiev; Mikhail Semenov; Irina Podlipnova
  9. Nested Optimal Transport Distances By Ruben Bontorno; Songyan Hou
  10. Reinforcement Learning-Based Market Making as a Stochastic Control on Non-Stationary Limit Order Book Dynamics By Rafael Zimmer; Oswaldo Luiz do Valle Costa
  11. Machine Learning Enhanced Multi-Factor Quantitative Trading: A Cross-Sectional Portfolio Optimization Approach with Bias Correction By Yimin Du
  12. Causal PDE-Control Models: A Structural Framework for Dynamic Portfolio Optimization By Alejandro Rodriguez Dominguez
  13. Predicting Market Troughs: A Machine Learning Approach with Causal Interpretation By Peilin Rao; Randall R. Rojas
  14. Deep Learning Option Pricing with Market Implied Volatility Surfaces By Lijie Ding; Egang Lu; Kin Cheung
  15. News Sentiment Embeddings for Stock Price Forecasting By Ayaan Qayyum
  16. Myopic Optimality: why reinforcement learning portfolio management strategies lose money By Yuming Ma
  17. Accelerated Portfolio Optimization and Option Pricing with Reinforcement Learning By Hadi Keramati; Samaneh Jazayeri
  18. Random Forests for Labor Market Analysis: Balancing Precision and Interpretability By Daniel Graeber; Lorenz Meister; Carsten Schröder; Sabine Zinn
  19. Beyond GARCH: Bayesian Neural Stochastic Volatility By Guo, Hongfei; Marín Díazaraque, Juan Miguel; Veiga, Helena
  20. Overparametrized models with posterior drift By Guillaume Coqueret; Martial Laguerre
  21. Transformers Beyond Order: A Chaos-Markov-Gaussian Framework for Short-Term Sentiment Forecasting of Any Financial OHLC timeseries Data By Arif Pathan
  22. Neural ARFIMA model for forecasting BRIC exchange rates with long memory under oil shocks and policy uncertainties By Tanujit Chakraborty; Donia Besher; Madhurima Panja; Shovon Sengupta
  23. P-CRE-DML: A Novel Approach for Causal Inference in Non-Linear Panel Data By Amarendra Sharma
  24. Forecasting Labor Markets with LSTNet: A Multi-Scale Deep Learning Approach By Adam Nelson-Archer; Aleia Sen; Meena Al Hasani; Sofia Davila; Jessica Le; Omar Abbouchi
  25. Governing Synthetic Data in the Financial Sector By Spears, Taylor C.; Hansen, Kristian Bondo; Xu, Ruowen; Millo, Yuval
  26. Deep Learning for Conditional Asset Pricing Models By Hongyi Liu
  27. Optimization Method of Multi-factor Investment Model Driven by Deep Learning for Risk Control By Ruisi Li; Xinhui Gu
  28. Integration of Wavelet Transform Convolution and Channel Attention with LSTM for Stock Price Prediction based Portfolio Allocation By Junjie Guo
  29. AI Agents for Economic Research By Anton Korinek
  30. Automated regime classification in multidimensional time series data using sliced Wasserstein k-means clustering By Luan, Qinmeng; Hamp, James
  31. Painting the market: generative diffusion models for financial limit order book simulation and forecasting By Alfred Backhouse; Kang Li; Jakob Foerster; Anisoara Calinescu; Stefan Zohren
  32. Context-Aware Language Models for Forecasting Market Impact from Sequences of Financial News By Ross Koval; Nicholas Andrews; Xifeng Yan
  33. Extreme-case Range Value-at-Risk under Increasing Failure Rate By Yuting Su; Taizhong Hu; Zhenfeng Zou
  34. Learning in Random Utility Models Via Online Decision Problems By Emerson Melo
  35. Epsilon-Minimax Solutions of Statistical Decision Problems By Andr\'es Aradillas Fern\'andez; Jos\'e Blanchet; Jos\'e Luis Montiel Olea; Chen Qiu; J\"org Stoye; Lezhi Tan

  1. By: Patrick J. Laub; Tu Pho; Bernard Wong
    Abstract: This paper introduces the Actuarial Neural Additive Model, an inherently interpretable deep learning model for general insurance pricing that offers fully transparent and interpretable results while retaining the strong predictive power of neural networks. This model assigns a dedicated neural network (or subnetwork) to each individual covariate and pairwise interaction term to independently learn its impact on the modeled output while implementing various architectural constraints to allow for essential interpretability (e.g. sparsity) and practical requirements (e.g. smoothness, monotonicity) in insurance applications. The development of our model is grounded in a solid foundation, where we establish a concrete definition of interpretability within the insurance context, complemented by a rigorous mathematical framework. Comparisons in terms of prediction accuracy are made with traditional actuarial and state-of-the-art machine learning methods using both synthetic and real insurance datasets. The results show that the proposed model outperforms other methods in most cases while offering complete transparency in its internal logic, underscoring the strong interpretability and predictive capability.
    Date: 2025–09
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2509.08467
  2. By: Yang Chen; Yueheng Jiang; Zhaozhao Ma; Yuchen Cao Jacky Keung; Kun Kuang; Leilei Gan; Yiquan Wu; Fei Wu
    Abstract: The inherent non-stationarity of financial markets and the complexity of multi-modal information pose significant challenges to existing quantitative trading models. Traditional methods relying on fixed structures and unimodal data struggle to adapt to market regime shifts, while large language model (LLM)-driven solutions - despite their multi-modal comprehension - suffer from static strategies and homogeneous expert designs, lacking dynamic adjustment and fine-grained decision mechanisms. To address these limitations, we propose MM-DREX: a Multimodal-driven, Dynamically-Routed EXpert framework based on large language models. MM-DREX explicitly decouples market state perception from strategy execution to enable adaptive sequential decision-making in non-stationary environments. Specifically, it (1) introduces a vision-language model (VLM)-powered dynamic router that jointly analyzes candlestick chart patterns and long-term temporal features to allocate real-time expert weights; (2) designs four heterogeneous trading experts (trend, reversal, breakout, positioning) generating specialized fine-grained sub-strategies; and (3) proposes an SFT-RL hybrid training paradigm to synergistically optimize the router's market classification capability and experts' risk-adjusted decision-making. Extensive experiments on multi-modal datasets spanning stocks, futures, and cryptocurrencies demonstrate that MM-DREX significantly outperforms 15 baselines (including state-of-the-art financial LLMs and deep reinforcement learning models) across key metrics: total return, Sharpe ratio, and maximum drawdown, validating its robustness and generalization. Additionally, an interpretability module traces routing logic and expert behavior in real time, providing an audit trail for strategy transparency.
    Date: 2025–09
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2509.05080
  3. By: Yiran Wan; Xinyu Ying; Shengzhen Xu
    Abstract: Straddle Option is a financial trading tool that explores volatility premiums in high-volatility markets without predicting price direction. Although deep reinforcement learning has emerged as a powerful approach to trading automation in financial markets, existing work mostly focused on predicting price trends and making trading decisions by combining multi-dimensional datasets like blogs and videos, which led to high computational costs and unstable performance in high-volatility markets. To tackle this challenge, we develop automated straddle option trading based on reinforcement learning and attention mechanisms to handle unpredictability in high-volatility markets. Firstly, we leverage the attention mechanisms in Transformer-DDQN through both self-attention with time series data and channel attention with multi-cycle information. Secondly, a novel reward function considering excess earnings is designed to focus on long-term profits and neglect short-term losses over a stop line. Thirdly, we identify the resistance levels to provide reference information when great uncertainty in price movements occurs with intensified battle between the buyers and sellers. Through extensive experiments on the Chinese stock, Brent crude oil, and Bitcoin markets, our attention-based Transformer-DDQN model exhibits the lowest maximum drawdown across all markets, and outperforms other models by 92.5\% in terms of the average return excluding the crude oil market due to relatively low fluctuation.
    Date: 2025–08
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2509.07987
  4. By: Feliks Ba\'nka (Warsaw University of Technology, Faculty of Electronics and Information Technology); Jaros{\l}aw A. Chudziak (Warsaw University of Technology)
    Abstract: In volatile financial markets, balancing risk and return remains a significant challenge. Traditional approaches often focus solely on equity allocation, overlooking the strategic advantages of options trading for dynamic risk hedging. This work presents DeltaHedge, a multi-agent framework that integrates options trading with AI-driven portfolio management. By combining advanced reinforcement learning techniques with an ensembled options-based hedging strategy, DeltaHedge enhances risk-adjusted returns and stabilizes portfolio performance across varying market conditions. Experimental results demonstrate that DeltaHedge outperforms traditional strategies and standalone models, underscoring its potential to transform practical portfolio management in complex financial environments. Building on these findings, this paper contributes to the fields of quantitative finance and AI-driven portfolio optimization by introducing a novel multi-agent system for integrating options trading strategies, addressing a gap in the existing literature.
    Date: 2025–09
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2509.12753
  5. By: Schmidt, Tobias; Lange, Kai-Robin; Reccius, Matthias; Müller, Henrik; Roos, Michael W. M.; Jentsch, Carsten
    Abstract: As interest in economic narratives has grown in recent years, so has the number of pipelines dedicated to extracting such narratives from texts. Pipelines often employ a mix of state-of-the-art natural language processing techniques, such as BERT, to tackle this task. While effective on foundational linguistic operations essential for narrative extraction, such models lack the deeper semantic understanding required to distinguish extracting economic narratives from merely conducting classic tasks like Semantic Role Labeling. Instead of relying on complex model pipelines, we evaluate the benefits of Large Language Models (LLMs) by analyzing a corpus of Wall Street Journal and New York Times newspaper articles about inflation. We apply a rigorous narrative definition and compare GPT 4o outputs to gold-standard narratives produced by expert annotators. Our results suggests that GPT-4o is capable of extracting valid economic narratives in a structured format, but still falls short of expert-level performance when handling complex documents and narratives. Given the novelty of LLMs in economic research, we also provide guidance for future work in economics and the social sciences that employs LLMs to pursue similar objectives.
    Keywords: Economic narratives, natural language processing, large language models
    JEL: C18 C55 C87 E70
    Date: 2025
    URL: https://d.repec.org/n?u=RePEc:zbw:rwirep:325494
  6. By: Michael Monoyios; Olivia Pricilia
    Abstract: We introduce a novel neural-network-based approach to learning the generating function $G(\cdot)$ of a functionally generated portfolio (FGP) from synthetic or real market data. In the neural network setting, the generating function is represented as $G_{\theta}(\cdot)$, where $\theta$ is an iterable neural network parameter vector, and $G_{\theta}(\cdot)$ is trained to maximise investment return relative to the market portfolio. We compare the performance of the Neural FGP approach against classical FGP benchmarks. FGPs provide a robust alternative to classical portfolio optimisation by bypassing the need to estimate drifts or covariances. The neural FGP framework extends this by introducing flexibility in the design of the generating function, enabling it to learn from market dynamics while preserving self-financing and pathwise decomposition properties.
    Date: 2025–06
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2506.19715
  7. By: Adrian Iulian Cristescu; Matteo Giordano
    Abstract: Predicting the probability of default (PD) of prospective loans is a critical objective for financial institutions. In recent years, machine learning (ML) algorithms have achieved remarkable success across a wide variety of prediction tasks; yet, they remain relatively underutilised in credit risk analysis. This paper highlights the opportunities that ML algorithms offer to this field by comparing the performance of five predictive models-Random Forests, Decision Trees, XGBoost, Gradient Boosting and AdaBoost-to the predominantly used logistic regression, over a benchmark dataset from Scheule et al. (Credit Risk Analytics: The R Companion). Our findings underscore the strengths and weaknesses of each method, providing valuable insights into the most effective ML algorithms for PD prediction in the context of loan portfolios.
    Date: 2025–06
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2506.19789
  8. By: Kasymkhan Khubiev; Mikhail Semenov; Irina Podlipnova
    Abstract: Deep Learning is evolving fast and integrates into various domains. Finance is a challenging field for deep learning, especially in the case of interpretable artificial intelligence (AI). Although classical approaches perform very well with natural language processing, computer vision, and forecasting, they are not perfect for the financial world, in which specialists use different metrics to evaluate model performance. We first introduce financially grounded loss functions derived from key quantitative finance metrics, including the Sharpe ratio, Profit-and-Loss (PnL), and Maximum Draw down. Additionally, we propose turnover regularization, a method that inherently constrains the turnover of generated positions within predefined limits. Our findings demonstrate that the proposed loss functions, in conjunction with turnover regularization, outperform the traditional mean squared error loss for return prediction tasks when evaluated using algorithmic trading metrics. The study shows that financially grounded metrics enhance predictive performance in trading strategies and portfolio optimization.
    Date: 2025–09
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2509.04541
  9. By: Ruben Bontorno; Songyan Hou
    Abstract: Simulating realistic financial time series is essential for stress testing, scenario generation, and decision-making under uncertainty. Despite advances in deep generative models, there is no consensus metric for their evaluation. We focus on generative AI for financial time series in decision-making applications and employ the nested optimal transport distance, a time-causal variant of optimal transport distance, which is robust to tasks such as hedging, optimal stopping, and reinforcement learning. Moreover, we propose a statistically consistent, naturally parallelizable algorithm for its computation, achieving substantial speedups over existing approaches.
    Date: 2025–09
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2509.06702
  10. By: Rafael Zimmer; Oswaldo Luiz do Valle Costa
    Abstract: Reinforcement Learning has emerged as a promising framework for developing adaptive and data-driven strategies, enabling market makers to optimize decision-making policies based on interactions with the limit order book environment. This paper explores the integration of a reinforcement learning agent in a market-making context, where the underlying market dynamics have been explicitly modeled to capture observed stylized facts of real markets, including clustered order arrival times, non-stationary spreads and return drifts, stochastic order quantities and price volatility. These mechanisms aim to enhance stability of the resulting control agent, and serve to incorporate domain-specific knowledge into the agent policy learning process. Our contributions include a practical implementation of a market making agent based on the Proximal-Policy Optimization (PPO) algorithm, alongside a comparative evaluation of the agent's performance under varying market conditions via a simulator-based environment. As evidenced by our analysis of the financial return and risk metrics when compared to a closed-form optimal solution, our results suggest that the reinforcement learning agent can effectively be used under non-stationary market conditions, and that the proposed simulator-based environment can serve as a valuable tool for training and pre-training reinforcement learning agents in market-making scenarios.
    Date: 2025–09
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2509.12456
  11. By: Yimin Du
    Abstract: This paper presents a comprehensive machine learning framework for quantitative trading that achieves superior risk-adjusted returns through systematic factor engineering, real-time computation optimization, and cross-sectional portfolio construction. Our approach integrates multi-factor alpha discovery with bias correction techniques, leveraging PyTorch-accelerated factor computation and advanced portfolio optimization. The system processes 500-1000 factors derived from open-source alpha101 extensions and proprietary market microstructure signals. Key innovations include tensor-based factor computation acceleration, geometric Brownian motion data augmentation, and cross-sectional neutralization strategies. Empirical validation on Chinese A-share markets (2010-2024) demonstrates annualized returns of $20\%$ with Sharpe ratios exceeding 2.0, significantly outperforming traditional approaches. Our analysis reveals the critical importance of bias correction in factor construction and the substantial impact of cross-sectional portfolio optimization on strategy performance. Code and experimental implementations are available at: https://github.com/initial-d/ml-quant-tr ading
    Date: 2025–06
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2507.07107
  12. By: Alejandro Rodriguez Dominguez
    Abstract: Classical portfolio models collapse under structural breaks, while modern machine-learning allocators adapt flexibly but often at the cost of transparency and interpretability. This paper introduces Causal PDE-Control Models (CPCMs), a unifying framework that integrates causal inference, nonlinear filtering, and forward-backward partial differential equations for dynamic portfolio optimization. The framework delivers three theoretical advances: (i) the existence of conditional risk-neutral measures under evolving information sets; (ii) a projection-divergence duality that quantifies the stability cost of departing from the causal driver manifold; and (iii) causal completeness, establishing that a finite driver span can capture all systematic premia. Classical methods such as Markowitz, CAPM, and Black-Litterman appear as degenerate cases, while reinforcement learning and deep-hedging policies emerge as unconstrained, symmetry-breaking approximations. Empirically, CPCM solvers implemented with physics-informed neural networks achieve higher Sharpe ratios, lower turnover, and more persistent premia than both econometric and machine-learning benchmarks, using a global equity panel with more than 300 candidate drivers. By reframing portfolio optimization around structural causality and PDE control, CPCMs provide a rigorous, interpretable, and computationally tractable foundation for robust asset allocation under nonstationary conditions.
    Date: 2025–09
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2509.09585
  13. By: Peilin Rao; Randall R. Rojas
    Abstract: This paper provides robust, new evidence on the causal drivers of market troughs. We demonstrate that conclusions about these triggers are critically sensitive to model specification, moving beyond restrictive linear models with a flexible DML average partial effect causal machine learning framework. Our robust estimates identify the volatility of options-implied risk appetite and market liquidity as key causal drivers, relationships misrepresented or obscured by simpler models. These findings provide high-frequency empirical support for intermediary asset pricing theories. This causal analysis is enabled by a high-performance nowcasting model that accurately identifies capitulation events in real-time.
    Date: 2025–09
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2509.05922
  14. By: Lijie Ding; Egang Lu; Kin Cheung
    Abstract: We present a deep learning framework for pricing options based on market-implied volatility surfaces. Using end-of-day S\&P 500 index options quotes from 2018-2023, we construct arbitrage-free volatility surfaces and generate training data for American puts and arithmetic Asian options using QuantLib. To address the high dimensionality of volatility surfaces, we employ a variational autoencoder (VAE) that compresses volatility surfaces across maturities and strikes into a 10-dimensional latent representation. We feed these latent variables, combined with option-specific inputs such as strike and maturity, into a multilayer perceptron to predict option prices. Our model is trained in stages: first to train the VAE for volatility surface compression and reconstruction, then options pricing mapping, and finally fine-tune the entire network end-to-end. The trained pricer achieves high accuracy across American and Asian options, with prediction errors concentrated primarily near long maturities and at-the-money strikes, where absolute bid-ask price differences are known to be large. Our method offers an efficient and scalable approach requiring only a single neural network forward pass and naturally improve with additional data. By bridging volatility surface modeling and option pricing in a unified framework, it provides a fast and flexible alternative to traditional numerical approaches for exotic options.
    Date: 2025–09
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2509.05911
  15. By: Ayaan Qayyum
    Abstract: This paper will discuss how headline data can be used to predict stock prices. The stock price in question is the SPDR S&P 500 ETF Trust, also known as SPY that tracks the performance of the largest 500 publicly traded corporations in the United States. A key focus is to use news headlines from the Wall Street Journal (WSJ) to predict the movement of stock prices on a daily timescale with OpenAI-based text embedding models used to create vector encodings of each headline with principal component analysis (PCA) to exact the key features. The challenge of this work is to capture the time-dependent and time-independent, nuanced impacts of news on stock prices while handling potential lag effects and market noise. Financial and economic data were collected to improve model performance; such sources include the U.S. Dollar Index (DXY) and Treasury Interest Yields. Over 390 machine-learning inference models were trained. The preliminary results show that headline data embeddings greatly benefit stock price prediction by at least 40% compared to training and optimizing a machine learning system without headline data embeddings.
    Date: 2025–06
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2507.01970
  16. By: Yuming Ma
    Abstract: Myopic optimization (MO) outperforms reinforcement learning (RL) in portfolio management: RL yields lower or negative returns, higher variance, larger costs, heavier CVaR, lower profitability, and greater model risk. We model execution/liquidation frictions with mark-to-market accounting. Using Malliavin calculus (Clark-Ocone/BEL), we derive policy gradients and risk shadow price, unifying HJB and KKT. This gives dual gap and convergence results: geometric MO vs. RL floors. We quantify phantom profit in RL via Malliavin policy-gradient contamination analysis and define a control-affects-dynamics (CAD) premium of RL indicating plausibly positive.
    Date: 2025–09
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2509.12764
  17. By: Hadi Keramati; Samaneh Jazayeri
    Abstract: We present a reinforcement learning (RL)-driven framework for optimizing block-preconditioner sizes in iterative solvers used in portfolio optimization and option pricing. The covariance matrix in portfolio optimization or the discretization of differential operators in option pricing models lead to large linear systems of the form $\mathbf{A}\textbf{x}=\textbf{b}$. Direct inversion of high-dimensional portfolio or fine-grid option pricing incurs a significant computational cost. Therefore, iterative methods are usually used for portfolios in real-world situations. Ill-conditioned systems, however, suffer from slow convergence. Traditional preconditioning techniques often require problem-specific parameter tuning. To overcome this limitation, we rely on RL to dynamically adjust the block-preconditioner sizes and accelerate iterative solver convergence. Evaluations on a suite of real-world portfolio optimization matrices demonstrate that our RL framework can be used to adjust preconditioning and significantly accelerate convergence and reduce computational cost. The proposed accelerated solver supports faster decision-making in dynamic portfolio allocation and real-time option pricing.
    Date: 2025–06
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2507.01972
  18. By: Daniel Graeber; Lorenz Meister; Carsten Schröder; Sabine Zinn
    Abstract: Machine learning is increasingly used in social science research, especially for prediction. However, the results are sometimes not as straight-forward to interpret compared to classic regression models. In this paper, we address this trade-off by comparing the predictive performance of random forests and logit regressions to analyze labor market vulnerabilities during the COVID-19 pandemic, and a global surrogate model to enhance our understanding of the complex dynamics. Our study shows that, especially in the presence of non-linearities and feature interactions, random forests outperform regressions both in predictive accuracy and interpretability, yielding policy-relevant insights on vulnerable groups affected by labor market disruptions
    Keywords: Machine learning, interpretability, labor market, random forests
    JEL: C45 C53 C25 J08 I18 C83 J21
    Date: 2025
    URL: https://d.repec.org/n?u=RePEc:diw:diwsop:diw_sp1230
  19. By: Guo, Hongfei; Marín Díazaraque, Juan Miguel; Veiga, Helena
    Abstract: Accurately forecasting volatility is central to risk management, portfolio allocation, and asset pricing. While high-frequency realised measures have been shown to improve predictive accuracy, their value is not uniform across markets or horizons. This paper introduces a class of Bayesian neural network stochastic volatility (NN-SV) models that combine the flexibility of machine learning with the structure of stochastic volatility models. The specifications incorporate realised variance, jump variation, and semivariance from daily and intraday data, and model uncertainty is addressed through a Bayesian stacking ensemble that adaptively aggregates predictive distributions. Using data from the DAX, FTSE 100, and S&P 500 indices, the models are evaluated against classical GARCH and parametric SV benchmarks. The results show that the predictive content of high-frequency measures is horizon- and market-specific. The Bayesian ensemble further enhances robustness by exploiting complementary model strengths. Overall, NN-SV models not only outperform established benchmarks in many settings but also provide new insights into market-specific drivers of volatility dynamics.
    Keywords: Ensemble forecasts; GARCH; Neural networks; Realised volatility; Stochastic volatility
    JEL: C11 C32 C45 C53 C58
    Date: 2025–09–16
    URL: https://d.repec.org/n?u=RePEc:cte:wsrepe:47944
  20. By: Guillaume Coqueret; Martial Laguerre
    Abstract: This paper investigates the impact of posterior drift on out-of-sample forecasting accuracy in overparametrized machine learning models. We document the loss in performance when the loadings of the data generating process change between the training and testing samples. This matters crucially in settings in which regime changes are likely to occur, for instance, in financial markets. Applied to equity premium forecasting, our results underline the sensitivity of a market timing strategy to sub-periods and to the bandwidth parameters that control the complexity of the model. For the average investor, we find that focusing on holding periods of 15 years can generate very heterogeneous returns, especially for small bandwidths. Large bandwidths yield much more consistent outcomes, but are far less appealing from a risk-adjusted return standpoint. All in all, our findings tend to recommend cautiousness when resorting to large linear models for stock market predictions.
    Date: 2025–06
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2506.23619
  21. By: Arif Pathan
    Abstract: Short-term sentiment forecasting in financial markets (e.g., stocks, indices) is challenging due to volatility, non-linearity, and noise in OHLC (Open, High, Low, Close) data. This paper introduces a novel CMG (Chaos-Markov-Gaussian) framework that integrates chaos theory, Markov property, and Gaussian processes to improve prediction accuracy. Chaos theory captures nonlinear dynamics; the Markov chain models regime shifts; Gaussian processes add probabilistic reasoning. We enhance the framework with transformer-based deep learning models to capture temporal patterns efficiently. The CMG Framework is designed for fast, resource-efficient, and accurate forecasting of any financial instrument's OHLC time series. Unlike traditional models that require heavy infrastructure and instrument-specific tuning, CMG reduces overhead and generalizes well. We evaluate the framework on market indices, forecasting sentiment for the next trading day's first quarter. A comparative study against statistical, ML, and DL baselines trained on the same dataset with no feature engineering shows CMG consistently outperforms in accuracy and efficiency, making it valuable for analysts and financial institutions.
    Date: 2025–06
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2506.17244
  22. By: Tanujit Chakraborty; Donia Besher; Madhurima Panja; Shovon Sengupta
    Abstract: Accurate forecasting of exchange rates remains a persistent challenge, particularly for emerging economies such as Brazil, Russia, India, and China (BRIC). These series exhibit long memory, nonlinearity, and non-stationarity properties that conventional time series models struggle to capture. Additionally, there exist several key drivers of exchange rate dynamics, including global economic policy uncertainty, US equity market volatility, US monetary policy uncertainty, oil price growth rates, and country-specific short-term interest rate differentials. These empirical complexities underscore the need for a flexible modeling framework that can jointly accommodate long memory, nonlinearity, and the influence of external drivers. To address these challenges, we propose a Neural AutoRegressive Fractionally Integrated Moving Average (NARFIMA) model that combines the long-memory representation of ARFIMA with the nonlinear learning capacity of neural networks, while flexibly incorporating exogenous causal variables. We establish theoretical properties of the model, including asymptotic stationarity of the NARFIMA process using Markov chains and nonlinear time series techniques. We quantify forecast uncertainty using conformal prediction intervals within the NARFIMA framework. Empirical results across six forecast horizons show that NARFIMA consistently outperforms various state-of-the-art statistical and machine learning models in forecasting BRIC exchange rates. These findings provide new insights for policymakers and market participants navigating volatile financial conditions. The \texttt{narfima} \textbf{R} package provides an implementation of our approach.
    Date: 2025–09
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2509.06697
  23. By: Amarendra Sharma
    Abstract: This paper introduces a novel Proxy-Enhanced Correlated Random Effects Double Machine Learning (P-CRE-DML) framework to estimate causal effects in panel data with non-linearities and unobserved heterogeneity. Combining Double Machine Learning (DML, Chernozhukov et al., 2018), Correlated Random Effects (CRE, Mundlak, 1978), and lagged variables (Arellano & Bond, 1991) and innovating within the CRE-DML framework (Chernozhukov et al., 2022; Clarke & Polselli, 2025; Fuhr & Papies, 2024), we apply P-CRE-DML to investigate the effect of social trust on GDP growth across 89 countries (2010-2020). We find positive and statistically significant relationship between social trust and economic growth. This aligns with prior findings on trust-growth relationship (e.g., Knack & Keefer, 1997). Furthermore, a Monte Carlo simulation demonstrates P-CRE-DML's advantage in terms of lower bias over CRE-DML and System GMM. P-CRE-DML offers a robust and flexible alternative for panel data causal inference, with applications beyond economic growth.
    Date: 2025–06
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2506.23297
  24. By: Adam Nelson-Archer; Aleia Sen; Meena Al Hasani; Sofia Davila; Jessica Le; Omar Abbouchi
    Abstract: We present a deep learning approach for forecasting short-term employment changes and assessing long-term industry health using labor market data from the U.S. Bureau of Labor Statistics. Our system leverages a Long- and Short-Term Time-series Network (LSTNet) to process multivariate time series data, including employment levels, wages, turnover rates, and job openings. The model outputs both 7-day employment forecasts and an interpretable Industry Employment Health Index (IEHI). Our approach outperforms baseline models across most sectors, particularly in stable industries, and demonstrates strong alignment between IEHI rankings and actual employment volatility. We discuss error patterns, sector-specific performance, and future directions for improving interpretability and generalization.
    Date: 2025–06
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2507.01979
  25. By: Spears, Taylor C. (University of Edinburgh); Hansen, Kristian Bondo; Xu, Ruowen; Millo, Yuval
    Abstract: Synthetic datasets, artificially generated to mimic real-world data while maintaining anonymization, have emerged as a promising technology in the financial sector, attracting support from regulators and market participants as a solution to data privacy and scarcity challenges limiting machine learning deployment. This paper argues that synthetic data's effects on financial markets depend critically on how these technologies are embedded within existing machine learning infrastructural ``stacks'' rather than on their intrinsic properties. We identify three key tensions that will determine whether adoption proves beneficial or harmful: (1) data circulability versus opacity, particularly the "double opacity" problem arising from stacked machine learning systems, (2) model-induced scattering versus model-induced herding in market participant behaviour, and (3) flattening versus deepening of data platform power. These tensions directly correspond to core regulatory priorities around model risk management, systemic risk, and competition policy. Using financial audit as a case study, we demonstrate how these tensions interact in practice and propose governance frameworks, including a synthetic data labelling regime to preserve contextual information when datasets cross organizational boundaries.
    Date: 2025–09–08
    URL: https://d.repec.org/n?u=RePEc:osf:socarx:ruxkh_v1
  26. By: Hongyi Liu
    Abstract: We propose a new pseudo-Siamese Network for Asset Pricing (SNAP) model, based on deep learning approaches, for conditional asset pricing. Our model allows for the deep alpha, deep beta and deep factor risk premia conditional on high dimensional observable information of financial characteristics and macroeconomic states, while storing the long-term dependency of the informative features through long short-term memory network. We apply this method to monthly U.S. stock returns from 1970-2019 and find that our pseudo-SNAP model outperforms the benchmark approaches in terms of out-of-sample prediction and out-of-sample Sharpe ratio. In addition, we also apply our method to calculate deep mispricing errors which we use to construct an arbitrage portfolio K-Means clustering. We find that the arbitrage portfolio has significant alphas.
    Date: 2025–09
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2509.04812
  27. By: Ruisi Li; Xinhui Gu
    Abstract: Propose a deep learning driven multi factor investment model optimization method for risk control. By constructing a deep learning model based on Long Short Term Memory (LSTM) and combining it with a multi factor investment model, we optimize factor selection and weight determination to enhance the model's adaptability and robustness to market changes. Empirical analysis shows that the LSTM model is significantly superior to the benchmark model in risk control indicators such as maximum retracement, Sharp ratio and value at risk (VaR), and shows strong adaptability and robustness in different market environments. Furthermore, the model is applied to the actual portfolio to optimize the asset allocation, which significantly improves the performance of the portfolio, provides investors with more scientific and accurate investment decision-making basis, and effectively balances the benefits and risks.
    Date: 2025–06
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2507.00332
  28. By: Junjie Guo
    Abstract: Portfolio allocation via stock price prediction is inherently difficult due to the notoriously low signal-to-noise ratio of stock time series. This paper proposes a method by integrating wavelet transform convolution and channel attention with LSTM to implement stock price prediction based portfolio allocation. Stock time series data first are processed by wavelet transform convolution to reduce the noise. Processed features are then reconstructed by channel attention. LSTM is utilized to predict the stock price using the final processed features. We construct a portfolio consists of four stocks with trading signals predicted by model. Experiments are conducted by evaluating the return, Sharpe ratio and max drawdown performance. The results indicate that our method achieves robust performance even during period of post-pandemic downward market.
    Date: 2025–06
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2507.01973
  29. By: Anton Korinek
    Abstract: The objective of this paper is to demystify AI agents - autonomous LLM-based systems that plan, use tools, and execute multi-step research tasks - and to provide hands-on instructions for economists to build their own, even if they do not have programming expertise. As AI has evolved from simple chatbots to reasoning models and now to autonomous agents, the main focus of this paper is to make these powerful tools accessible to all researchers. Through working examples and step-by-step code, it shows how economists can create agents that autonomously conduct literature reviews across myriads of sources, write and debug econometric code, fetch and analyze economic data, and coordinate complex research workflows. The paper demonstrates that by "vibe coding" (programming through natural language) and building on modern agentic frameworks like LangGraph, any economist can build sophisticated research assistants and other autonomous tools in minutes. By providing complete, working implementations alongside conceptual frameworks, this guide demonstrates how to employ AI agents in every stage of the research process, from initial investigation to final analysis.
    JEL: A11 B41 C63
    Date: 2025–09
    URL: https://d.repec.org/n?u=RePEc:nbr:nberwo:34202
  30. By: Luan, Qinmeng; Hamp, James
    Abstract: Recent work has proposed Wasserstein k-means (Wk-means) clustering as a powerful method to classify regimes in time series data, and one-dimensional asset returns in particular. In this paper, we begin by studying in detail the behaviour of the Wasserstein k-means clustering algorithm applied to synthetic one-dimensional time series data. We extend the previous work by studying, in detail, the dynamics of the clustering algorithm and how varying the hyperparameters impacts the performance over different random initialisations. We compute simple metrics that we find to be useful in identifying high-quality clusterings. We then extend the technique of Wasserstein k-means clustering to multidimensional time series data by approximating the multidimensional Wasserstein distance as a sliced Wasserstein distance, resulting in a method we call 'sliced Wasserstein k-means (sWk-means) clustering'. We apply the sWk-means clustering method to the problem of automated regime classification in multidimensional time series data, using synthetic data to demonstrate the validity and effectiveness of the approach. Finally, we show that the sWk-means method is able to identify distinct market regimes in real multidimensional financial time series, using publicly available foreign exchange spot rate data as a case study. We conclude with remarks about some limitations of our approach and potential complementary or alternative approaches.
    Keywords: Wasserstein metric; market regimes; regime classification; time series; unsupervised learning
    JEL: C14 C63
    Date: 2025–08–29
    URL: https://d.repec.org/n?u=RePEc:ehl:lserod:129537
  31. By: Alfred Backhouse; Kang Li; Jakob Foerster; Anisoara Calinescu; Stefan Zohren
    Abstract: Simulating limit order books (LOBs) has important applications across forecasting and backtesting for financial market data. However, deep generative models struggle in this context due to the high noise and complexity of the data. Previous work uses autoregressive models, although these experience error accumulation over longer-time sequences. We introduce a novel approach, converting LOB data into a structured image format, and applying diffusion models with inpainting to generate future LOB states. This method leverages spatio-temporal inductive biases in the order book and enables parallel generation of long sequences overcoming issues with error accumulation. We also publicly contribute to LOB-Bench, the industry benchmark for LOB generative models, to allow fair comparison between models using Level-2 and Level-3 order book data (with or without message level data respectively). We show that our model achieves state-of-the-art performance on LOB-Bench, despite using lower fidelity data as input. We also show that our method prioritises coherent global structures over local, high-fidelity details, providing significant improvements over existing methods on certain metrics. Overall, our method lays a strong foundation for future research into generative diffusion approaches to LOB modelling.
    Date: 2025–09
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2509.05107
  32. By: Ross Koval; Nicholas Andrews; Xifeng Yan
    Abstract: Financial news plays a critical role in the information diffusion process in financial markets and is a known driver of stock prices. However, the information in each news article is not necessarily self-contained, often requiring a broader understanding of the historical news coverage for accurate interpretation. Further, identifying and incorporating the most relevant contextual information presents significant challenges. In this work, we explore the value of historical context in the ability of large language models to understand the market impact of financial news. We find that historical context provides a consistent and significant improvement in performance across methods and time horizons. To this end, we propose an efficient and effective contextualization method that uses a large LM to process the main article, while a small LM encodes the historical context into concise summary embeddings that are then aligned with the large model's representation space. We explore the behavior of the model through multiple qualitative and quantitative interpretability tests and reveal insights into the value of contextualization. Finally, we demonstrate that the value of historical context in model predictions has real-world applications, translating to substantial improvements in simulated investment performance.
    Date: 2025–09
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2509.12519
  33. By: Yuting Su; Taizhong Hu; Zhenfeng Zou
    Abstract: The extreme cases of risk measures, when considered within the context of distributional ambiguity, provide significant guidance for practitioners specializing in risk management of quantitative finance and insurance. In contrast to the findings of preceding studies, we focus on the study of extreme-case risk measure under distributional ambiguity with the property of increasing failure rate (IFR). The extreme-case range Value-at-Risk under distributional uncertainty, consisting of given mean and/or variance of distributions with IFR, is provided. The specific characteristics of extreme-case distributions under these constraints have been characterized, a crucial step for numerical simulations. We then apply our main results to stop-loss and limited loss random variables under distributional uncertainty with IFR.
    Date: 2025–06
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2506.23073
  34. By: Emerson Melo
    Abstract: This paper examines the Random Utility Model (RUM) in repeated stochastic choice settings where decision-makers lack full information about payoffs. We propose a gradient-based learning algorithm that embeds RUM into an online decision-making framework. Our analysis establishes Hannan consistency for a broad class of RUMs, meaning the average regret relative to the best fixed action in hindsight vanishes over time. We also show that our algorithm is equivalent to the Follow-The-Regularized-Leader (FTRL) method, offering an economically grounded approach to online optimization. Applications include modeling recency bias and characterizing coarse correlated equilibria in normal-form games
    Date: 2025–06
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2506.16030
  35. By: Andr\'es Aradillas Fern\'andez; Jos\'e Blanchet; Jos\'e Luis Montiel Olea; Chen Qiu; J\"org Stoye; Lezhi Tan
    Abstract: A decision rule is epsilon-minimax if it is minimax up to an additive factor epsilon. We present an algorithm for provably obtaining epsilon-minimax solutions of statistical decision problems. We are interested in problems where the statistician chooses randomly among I decision rules. The minimax solution of these problems admits a convex programming representation over the (I-1)-simplex. Our suggested algorithm is a well-known mirror subgradient descent routine, designed to approximately solve the convex optimization problem that defines the minimax decision rule. This iterative routine is known in the computer science literature as the hedge algorithm and it is used in algorithmic game theory as a practical tool to find approximate solutions of two-person zero-sum games. We apply the suggested algorithm to different minimax problems in the econometrics literature. An empirical application to the problem of optimally selecting sites to maximize the external validity of an experimental policy evaluation illustrates the usefulness of the suggested procedure.
    Date: 2025–09
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2509.08107

This nep-cmp issue is ©2025 by Stan Miles. It is provided as is without any express or implied warranty. It may be freely redistributed in whole or in part for any purpose. If distributed in part, please include this notice.
General information on the NEP project can be found at https://nep.repec.org. For comments please write to the director of NEP, Marco Novarese at <director@nep.repec.org>. Put “NEP” in the subject, otherwise your mail may be rejected.
NEP’s infrastructure is sponsored by the School of Economics and Finance of Massey University in New Zealand.