|
on Computational Economics |
By: | Paulo Andr\'e Lima de Castro |
Abstract: | Autonomous trading strategies have been a subject of research within the field of artificial intelligence (AI) for aconsiderable period. Various AI techniques have been explored to develop autonomous agents capable of trading financial assets. These approaches encompass traditional methods such as neural networks, fuzzy logic, and reinforcement learning, as well as more recent advancements, including deep neural networks and deep reinforcement learning. Many developers report success in creating strategies that exhibit strong performance during simulations using historical price data, a process commonly referred to as backtesting. However, when these strategies are deployed in real markets, their performance often deteriorates, particularly in terms of risk-adjusted returns. In this study, we propose an AI-based strategy inspired by a classical investment paradigm: Value Investing. Financial AI models are highly susceptible to lookahead bias and other forms of bias that can significantly inflate performance in backtesting compared to live trading conditions. To address this issue, we conducted a series of computational simulations while controlling for these biases, thereby reducing the risk of overfitting. Our results indicate that the proposed approach outperforms major Brazilian market benchmarks. Moreover, the strategy, named AlphaX, demonstrated superior performance relative to widely used technical indicators such as the Relative Strength Index (RSI) and Money Flow Index (MFI), with statistically significant results. Finally, we discuss several open challenges and highlight emerging technologies in qualitative analysis that may contribute to the development of a comprehensive AI-based Value Investing framework in the future |
Date: | 2025–08 |
URL: | https://d.repec.org/n?u=RePEc:arx:papers:2508.13429 |
By: | Simon Hatzesberger; Iris Nonneman |
Abstract: | This article demonstrates the transformative impact of Generative AI (GenAI) on actuarial science, illustrated by four implemented case studies. It begins with a historical overview of AI, tracing its evolution from early neural networks to modern GenAI technologies. The first case study shows how Large Language Models (LLMs) improve claims cost prediction by deriving significant features from unstructured textual data, significantly reducing prediction errors in the underlying machine learning task. In the second case study, we explore the automation of market comparisons using the GenAI concept of Retrieval-Augmented Generation to identify and process relevant information from documents. A third case study highlights the capabilities of fine-tuned vision-enabled LLMs in classifying car damage types and extracting contextual information. The fourth case study presents a multi-agent system that autonomously analyzes data from a given dataset and generates a corresponding report detailing the key findings. In addition to these case studies, we outline further potential applications of GenAI in the insurance industry, such as the automation of claims processing and fraud detection, and the verification of document compliance with internal or external policies. Finally, we discuss challenges and considerations associated with the use of GenAI, covering regulatory issues, ethical concerns, and technical limitations, among others. |
Date: | 2025–06 |
URL: | https://d.repec.org/n?u=RePEc:arx:papers:2506.18942 |
By: | Gorjian, Mahshid |
Abstract: | It is becoming increasingly important to predict property prices to mitigate investment risk, establish policies, and preserve market stability. To determine the practical utility and anticipated efficacy of the sophisticated statistical and machine learning models that have emerged, a comparative analysis is required. The purpose of this systematic study is to assess the predictive effectiveness and interpretability of hedonic regression and complex machine learning models in the estimation of housing prices in a wide range of foreign scenarios. In May 2024, a thorough search was conducted in Scopus, Google Scholar, and Web of Science. The search terms included "hedonic pricing models, " "machine learning, " and "housing price prediction, " in addition to others. The inclusion criteria required the utilization of empirical research published after 2000, a comparison of at least two predictive models, and reliable transaction data. Research that utilized non-empirical methodologies or web- scraped prices was excluded. Twenty-three investigations met the eligibility criteria. The evaluation was conducted in accordance with the reporting criteria of PRISMA 2020. Random Forest was the most frequently employed and consistently high-performing model, being selected in 14 of 23 studies and regarded as exceptional in five. Despite their lack of precision, hedonic regression models provided critical explanatory insights into critical variables, such as proximity to urban centers, property characteristics, and location. The integration of hedonic and machine learning models improved the interpretability and accuracy of the predicted results. Many of the studies included in this review were longitudinal, covered a diverse range of international contexts (specifically, Asia, Europe, America, and Australia), and demonstrated a rise in research output beyond 2020. Even though hedonic models retain a significant amount of explanatory power, the precision of home price predictions is improved by machine learning, particularly Random Forest and neural networks. The optimal results for researchers, real estate professionals, and policymakers who aim to improve market transparency and enlighten effective policy decisions are achieved through the seamless integration of these techniques. |
Keywords: | housing price prediction; machine learning; hedonic price model; Random Forest; real estate valuation; artificial neural networks; systematic review; property market analysis |
JEL: | C00 C01 C10 |
Date: | 2025 |
URL: | https://d.repec.org/n?u=RePEc:pra:mprapa:125676 |
By: | Linh Nguyen; Marcel Boersma; Erman Acar |
Abstract: | Fraudulent activity in the financial industry costs billions annually. Detecting fraud, therefore, is an essential yet technically challenging task that requires carefully analyzing large volumes of data. While machine learning (ML) approaches seem like a viable solution, applying them successfully is not so easy due to two main challenges: (1) the sparsely labeled data, which makes the training of such approaches challenging (with inherent labeling costs), and (2) lack of explainability for the flagged items posed by the opacity of ML models, that is often required by business regulations. This article proposes SAGE-FIN, a semi-supervised graph neural network (GNN) based approach with Granger causal explanations for Financial Interaction Networks. SAGE-FIN learns to flag fraudulent items based on weakly labeled (or unlabelled) data points. To adhere to regulatory requirements, the flagged items are explained by highlighting related items in the network using Granger causality. We empirically validate the favorable performance of SAGE-FIN on a real-world dataset, Bipartite Edge-And-Node Attributed financial network (Elliptic++), with Granger-causal explanations for the identified fraudulent items without any prior assumption on the network structure. |
Date: | 2025–06 |
URL: | https://d.repec.org/n?u=RePEc:arx:papers:2507.01980 |
By: | Jakub Micha\'nk\'ow |
Abstract: | This study evaluates deep neural networks for forecasting probability distributions of financial returns. 1D convolutional neural networks (CNN) and Long Short-Term Memory (LSTM) architectures are used to forecast parameters of three probability distributions: Normal, Student's t, and skewed Student's t. Using custom negative log-likelihood loss functions, distribution parameters are optimized directly. The models are tested on six major equity indices (S\&P 500, BOVESPA, DAX, WIG, Nikkei 225, and KOSPI) using probabilistic evaluation metrics including Log Predictive Score (LPS), Continuous Ranked Probability Score (CRPS), and Probability Integral Transform (PIT). Results show that deep learning models provide accurate distributional forecasts and perform competitively with classical GARCH models for Value-at-Risk estimation. The LSTM with skewed Student's t distribution performs best across multiple evaluation criteria, capturing both heavy tails and asymmetry in financial returns. This work shows that deep neural networks are viable alternatives to traditional econometric models for financial risk assessment and portfolio management. |
Date: | 2025–08 |
URL: | https://d.repec.org/n?u=RePEc:arx:papers:2508.18921 |
By: | Shrenik Jadhav; Birva Sevak; Srijita Das; Akhtar Hussain; Wencong Su; Van-Hai Bui |
Abstract: | Peer-to-peer (P2P) trading is increasingly recognized as a key mechanism for decentralized market regulation, yet existing approaches often lack robust frameworks to ensure fairness. This paper presents FairMarket-RL, a novel hybrid framework that combines Large Language Models (LLMs) with Reinforcement Learning (RL) to enable fairness-aware trading agents. In a simulated P2P microgrid with multiple sellers and buyers, the LLM acts as a real-time fairness critic, evaluating each trading episode using two metrics: Fairness-To-Buyer (FTB) and Fairness-Between-Sellers (FBS). These fairness scores are integrated into agent rewards through scheduled {\lambda}-coefficients, forming an adaptive LLM-guided reward shaping loop that replaces brittle, rule-based fairness constraints. Agents are trained using Independent Proximal Policy Optimization (IPPO) and achieve equitable outcomes, fulfilling over 90% of buyer demand, maintaining fair seller margins, and consistently reaching FTB and FBS scores above 0.80. The training process demonstrates that fairness feedback improves convergence, reduces buyer shortfalls, and narrows profit disparities between sellers. With its language-based critic, the framework scales naturally, and its extension to a large power distribution system with household prosumers illustrates its practical applicability. FairMarket-RL thus offers a scalable, equity-driven solution for autonomous trading in decentralized energy systems. |
Date: | 2025–06 |
URL: | https://d.repec.org/n?u=RePEc:arx:papers:2506.22708 |
By: | Chi-Sheng Chen; Xinyu Zhang; Ya-Chuan Chen |
Abstract: | We propose a hybrid quantum-classical reinforcement learning framework for sector rotation in the Taiwan stock market. Our system employs Proximal Policy Optimization (PPO) as the backbone algorithm and integrates both classical architectures (LSTM, Transformer) and quantum-enhanced models (QNN, QRWKV, QASA) as policy and value networks. An automated feature engineering pipeline extracts financial indicators from capital share data to ensure consistent model input across all configurations. Empirical backtesting reveals a key finding: although quantum-enhanced models consistently achieve higher training rewards, they underperform classical models in real-world investment metrics such as cumulative return and Sharpe ratio. This discrepancy highlights a core challenge in applying reinforcement learning to financial domains -- namely, the mismatch between proxy reward signals and true investment objectives. Our analysis suggests that current reward designs may incentivize overfitting to short-term volatility rather than optimizing risk-adjusted returns. This issue is compounded by the inherent expressiveness and optimization instability of quantum circuits under Noisy Intermediate-Scale Quantum (NISQ) constraints. We discuss the implications of this reward-performance gap and propose directions for future improvement, including reward shaping, model regularization, and validation-based early stopping. Our work offers a reproducible benchmark and critical insights into the practical challenges of deploying quantum reinforcement learning in real-world finance. |
Date: | 2025–06 |
URL: | https://d.repec.org/n?u=RePEc:arx:papers:2506.20930 |
By: | Nicholas Gray; Finn Lattimore; Kate McLoughlin; Callan Windsor |
Abstract: | In a world of increasing policy uncertainty, central banks are relying more on soft information sources to complement traditional economic statistics and model-based forecasts. One valuable source of soft information comes from intelligence gathered through central bank liaison programs -- structured programs in which central bank staff regularly talk with firms to gather insights. This paper introduces a new text analytics and retrieval tool that efficiently processes, organises, and analyses liaison intelligence gathered from firms using modern natural language processing techniques. The textual dataset spans 25 years, integrates new information as soon as it becomes available, and covers a wide range of business sizes and industries. The tool uses both traditional text analysis techniques and powerful language models to provide analysts and researchers with three key capabilities: (1) quickly querying the entire history of business liaison meeting notes; (2) zooming in on particular topics to examine their frequency (topic exposure) and analysing the associated tone and uncertainty of the discussion; and (3) extracting precise numerical values from the text, such as firms' reported figures for wages and prices growth. We demonstrate how these capabilities are useful for assessing economic conditions by generating text-based indicators of wages growth and incorporating them into a nowcasting model. We find that adding these text-based features to current best-in-class predictive models, combined with the use of machine learning methods designed to handle many predictors, significantly improves the performance of nowcasts for wages growth. Predictive gains are driven by a small number of features, indicating a sparse signal in contrast to other predictive problems in macroeconomics, where the signal is typically dense. |
Date: | 2025–06 |
URL: | https://d.repec.org/n?u=RePEc:arx:papers:2506.18505 |
By: | Umberto Collodel |
Abstract: | This paper develops a novel method to simulate financial market reactions to European Central Bank (ECB) press conferences using a Large Language Model (LLM). We create a behavioral, agent-based simulation of 30 synthetic traders, each with distinct risk preferences, cognitive biases, and interpretive styles. These agents forecast Euro interest rate swap levels at 3-month, 2-year, and 10-year maturities, with the variation across forecasts serving as a measure of market uncertainty or disagreement. We evaluate three prompting strategies, naive, few-shot (enriched with historical data), and an advanced iterative 'LLM-as-a-Judge' framework, to assess the effect of prompt design on predictive performance. Even the naive approach generates a strong correlation (roughly 0.5) between synthetic disagreement and actual market outcomes, particularly for longer-term maturities. The LLM-as-a-Judge framework further improves accuracy at the first iteration. These results demonstrate that LLM-driven simulations can capture interpretive uncertainty beyond traditional measures, providing central banks with a practical tool to anticipate market reactions, refine communication strategies, and enhance financial stability. |
Date: | 2025–08 |
URL: | https://d.repec.org/n?u=RePEc:arx:papers:2508.13635 |
By: | Fiona Xiao Jingyi; Lili Liu |
Abstract: | Forecasting central bank policy decisions remains a persistent challenge for investors, financial institutions, and policymakers due to the wide-reaching impact of monetary actions. In particular, anticipating shifts in the U.S. federal funds rate is vital for risk management and trading strategies. Traditional methods relying only on structured macroeconomic indicators often fall short in capturing the forward-looking cues embedded in central bank communications. This study examines whether predictive accuracy can be enhanced by integrating structured data with unstructured textual signals from Federal Reserve communications. We adopt a multi-modal framework, comparing traditional machine learning models, transformer-based language models, and deep learning architectures in both unimodal and hybrid settings. Our results show that hybrid models consistently outperform unimodal baselines. The best performance is achieved by combining TF-IDF features of FOMC texts with economic indicators in an XGBoost classifier, reaching a test AUC of 0.83. FinBERT-based sentiment features marginally improve ranking but perform worse in classification, especially under class imbalance. SHAP analysis reveals that sparse, interpretable features align more closely with policy-relevant signals. These findings underscore the importance of integrating textual and structured signals transparently. For monetary policy forecasting, simpler hybrid models can offer both accuracy and interpretability, delivering actionable insights for researchers and decision-makers. |
Date: | 2025–06 |
URL: | https://d.repec.org/n?u=RePEc:arx:papers:2506.22763 |
By: | Kai-Robin Lange; Tobias Schmidt; Matthias Reccius; Henrik M\"uller; Michael Roos; Carsten Jentsch |
Abstract: | With rapidly evolving media narratives, it has become increasingly critical to not just extract narratives from a given corpus but rather investigate, how they develop over time. While popular narrative extraction methods such as Large Language Models do well in capturing typical narrative elements or even the complex structure of a narrative, applying them to an entire corpus comes with obstacles, such as a high financial or computational cost. We propose a combination of the language understanding capabilities of Large Language Models with the large scale applicability of topic models to dynamically model narrative shifts across time using the Narrative Policy Framework. We apply a topic model and a corresponding change point detection method to find changes that concern a specific topic of interest. Using this model, we filter our corpus for documents that are particularly representative of that change and feed them into a Large Language Model that interprets the change that happened in an automated fashion and distinguishes between content and narrative shifts. We employ our pipeline on a corpus of The Wall Street Journal news paper articles from 2009 to 2023. Our findings indicate that a Large Language Model can efficiently extract a narrative shift if one exists at a given point in time, but does not perform as well when having to decide whether a shift in content or a narrative shift took place. |
Date: | 2025–06 |
URL: | https://d.repec.org/n?u=RePEc:arx:papers:2506.20269 |
By: | Mateusz Wilinski; Juho Kanniainen |
Abstract: | In this work we show how generative tools, which were successfully applied to limit order book data, can be utilized for the task of imitating trading agents. To this end, we propose a modified generative architecture based on the state-space model, and apply it to limit order book data with identified investors. The model is trained on synthetic data, generated from a heterogeneous agent-based model. Finally, we compare model's predicted distribution over different aspects of investors' actions, with the ground truths known from the agent-based model. |
Date: | 2025–08 |
URL: | https://d.repec.org/n?u=RePEc:arx:papers:2509.00982 |
By: | Alper Deniz Karakas |
Abstract: | This study examines the long-term economic impact of the colonial Mita system in Peru, building on Melissa Dell's foundational work on the enduring effects of forced labor institutions. The Mita, imposed by the Spanish colonial authorities from 1573 to 1812, required indigenous communities within a designated boundary to supply labor to mines, primarily near Potosi. Dell's original regression discontinuity design (RDD) analysis, leveraging the Mita boundary to estimate the Mita's legacy on modern economic outcomes, indicates that regions subjected to the Mita exhibit lower household consumption levels and higher rates of child stunting. In this paper, I replicate Dell's results and extend this analysis. I apply Double Machine Learning (DML) methods--the Partially Linear Regression (PLR) model and the Interactive Regression Model (IRM)--to further investigate the Mita's effects. DML allows for the inclusion of high-dimensional covariates and enables more flexible, non-linear modeling of treatment effects, potentially capturing complex relationships that a polynomial-based approach may overlook. While the PLR model provides some additional flexibility, the IRM model allows for fully heterogeneous treatment effects, offering a nuanced perspective on the Mita's impact across regions and district characteristics. My findings suggest that the Mita's economic legacy is more substantial and spatially heterogeneous than originally estimated. The IRM results reveal that proximity to Potosi and other district-specific factors intensify the Mita's adverse impact, suggesting a deeper persistence of regional economic inequality. These findings underscore that machine learning addresses the realistic non-linearity present in complex, real-world systems. By modeling hypothetical counterfactuals more accurately, DML enhances my ability to estimate the true causal impact of historical interventions. |
Date: | 2025–06 |
URL: | https://d.repec.org/n?u=RePEc:arx:papers:2506.18947 |
By: | Ahmet Umur \"Ozsoy |
Abstract: | We reinterpret and propose a framework for pricing path-dependent financial derivatives by estimating the full distribution of payoffs using Distributional Reinforcement Learning (DistRL). Unlike traditional methods that focus on expected option value, our approach models the entire conditional distribution of payoffs, allowing for risk-aware pricing, tail-risk estimation, and enhanced uncertainty quantification. We demonstrate the efficacy of this method on Asian options, using quantile-based value function approximators. |
Date: | 2025–07 |
URL: | https://d.repec.org/n?u=RePEc:arx:papers:2507.12657 |
By: | Irina G. Tanashkina; Alexey S. Tanashkin; Alexander S. Maksimchuik; Anna Yu. Poshivailo |
Abstract: | In this article, we review modern approaches to building interpretable models of property markets using machine learning on the base of mass valuation of property in the Primorye region, Russia. The researcher, lacking expertise in this topic, encounters numerous difficulties in the effort to build a good model. The main source of this is the huge difference between noisy real market data and ideal data which is very common in all types of tutorials on machine learning. This paper covers all stages of modeling: the collection of initial data, identification of outliers, the search and analysis of patterns in the data, the formation and final choice of price factors, the building of the model, and the evaluation of its efficiency. For each stage, we highlight potential issues and describe sound methods for overcoming emerging difficulties on actual examples. We show that the combination of classical linear regression with interpolation methods of geostatistics allows to build an effective model for land parcels. For flats, when many objects are attributed to one spatial point the application of geostatistical methods is difficult. Therefore we suggest linear regression with automatic generation and selection of additional rules on the base of decision trees, so called the RuleFit method. Thus we show, that despite such a strong restriction as the requirement of interpretability which is important in practical aspects, for example, legal matters, it is still possible to build effective models of real property markets. |
Date: | 2025–06 |
URL: | https://d.repec.org/n?u=RePEc:arx:papers:2506.15723 |
By: | Boris Kriuk; Logic Ng; Zarif Al Hossain |
Abstract: | Support and resistance (SR) levels are central to technical analysis, guiding traders in entry, exit, and risk management. Despite widespread use, traditional SR identification methods often fail to adapt to the complexities of modern, volatile markets. Recent research has introduced machine learning techniques to address the following challenges, yet most focus on price prediction rather than structural level identification. This paper presents DeepSupp, a new deep learning approach for detecting financial support levels using multi-head attention mechanisms to analyze spatial correlations and market microstructure relationships. DeepSupp integrates advanced feature engineering, constructing dynamic correlation matrices that capture evolving market relationships, and employs an attention-based autoencoder for robust representation learning. The final support levels are extracted through unsupervised clustering, leveraging DBSCAN to identify significant price thresholds. Comprehensive evaluations on S&P 500 tickers demonstrate that DeepSupp outperforms six baseline methods, achieving state-of-the-art performance across six financial metrics, including essential support accuracy and market regime sensitivity. With consistent results across diverse market conditions, DeepSupp addresses critical gaps in SR level detection, offering a scalable and reliable solution for modern financial analysis. Our approach highlights the potential of attention-based architectures to uncover nuanced market patterns and improve technical trading strategies. |
Date: | 2025–06 |
URL: | https://d.repec.org/n?u=RePEc:arx:papers:2507.01971 |
By: | Francis J. DiTraglia; Laura Liu |
Abstract: | This paper proposes a simple, novel, and fully-Bayesian approach for causal inference in partially linear models with high-dimensional control variables. Off-the-shelf machine learning methods can introduce biases in the causal parameter known as regularization-induced confounding. To address this, we propose a Bayesian Double Machine Learning (BDML) method, which modifies a standard Bayesian multivariate regression model and recovers the causal effect of interest from the reduced-form covariance matrix. Our BDML is related to the burgeoning frequentist literature on DML while addressing its limitations in finite-sample inference. Moreover, the BDML is based on a fully generative probability model in the DML context, adhering to the likelihood principle. We show that in high dimensional setups the naive estimator implicitly assumes no selection on observables--unlike our BDML. The BDML exhibits lower asymptotic bias and achieves asymptotic normality and semiparametric efficiency as established by a Bernstein-von Mises theorem, thereby ensuring robustness to misspecification. In simulations, our BDML achieves lower RMSE, better frequentist coverage, and shorter confidence interval width than alternatives from the literature, both Bayesian and frequentist. |
Date: | 2025–08 |
URL: | https://d.repec.org/n?u=RePEc:arx:papers:2508.12688 |
By: | Jonathan Garita-Garita (Department of Economic Research, Central Bank of Costa Rica); César Ulate-Sancho (Department of Economic Research, Central Bank of Costa Rica) |
Abstract: | This paper offers a daily-frequency analysis and short-term forecasting of Costa Rica’s foreign currency market using deep neural network algorithms. These algo-rithms efficiently integrates multiple high-frequency data to capture trends, seasonal patterns, and daily movements in the exchange rate from 2017 to March 2025. The results indicate that these models excels in predicting the observed exchange rate up to five days in advance, outperforming traditional time series forecasting methods in terms of accuracy. *** Resumen: Este artículo realiza un análisis de alta frecuencia del mercado de divisas de Costa Rica utilizando algoritmos de redes neuronales profundas. Se emplean datos diarios de acceso público de MONEX desde 2017 hasta marzo de 2025 para identificar quiebres de tendencia, patrones estacionales y la importancia relativa de las variables explicativas que determinan los movimientos diarios del tipo de cambio en MONEX. El modelo calibrado muestra una alta precisión para comprender la información histórica y realizar proyecciones del tipo de cambio a cinco días. Los resultados sugieren que los movimientos observados del tipo de cambio en 2024 están alineados con su tendencia y que existen factores estacionales significativos que influyen en el tipo de cambio a lo largo del año. |
Keywords: | Exchange Rate, Forecast, Deep Neural Network, Tipo de cambio, Pronóstico, Redes neuronales profundas |
JEL: | C45 C53 F31 O24 |
Date: | 2025–07 |
URL: | https://d.repec.org/n?u=RePEc:apk:doctra:2505 |
By: | Pesce, Simone; Errico, Marco; Pollio, Luigi |
Abstract: | Firms respond heterogeneously to aggregate fluctuations, yet standard linear models impose restrictive assumptions on firm sensitivities. Applying the Generalized Random Forest to U.S. firm-level data, we document strong nonlinearities in how firm characteristics shape responses to macroeconomic shocks. We show that nonlinearities significantly lower aggregate esponses, leading linear models to overestimate the economy’s sensitivity to shocks by up to 1.7 percentage points. We also find that larger firms, which carry disproportionate economic weight, exhibit lower sensitivities, leading to a median reduction in aggregate economic sensitivity of 52%. Our results highlight the importance of accounting for nonlinearities and firm heterogeneity when analyzing macroeconomic fluctuations and the transmission of aggregate shocks. JEL Classification: D22, E32, C14, E5 |
Keywords: | business cycle, firm sensitivity, monetary policy, oil shock, uncertainty |
Date: | 2025–09 |
URL: | https://d.repec.org/n?u=RePEc:ecb:ecbwps:20253107 |
By: | Yating Ru (Asian Development Bank); Elizabeth Tennant (Cornell University); David Matteson (Cornell University); Christopher Barrett (Cornell University) |
Abstract: | Recent studies harnessing geospatial big data and machine learning have significantly advanced poverty mapping, enabling granular and timely welfare estimates in traditionally data scarce regions. While much of the existing research has focused on overall out-of-sample predictive performance, there is a lack of understanding regarding where such models underperform and whether key spatial relationships might vary across places. This study investigates spatial heterogeneity in machine learning-based poverty mapping, testing whether spatial regression and machine learning techniques produce more unbiased predictions. We find that extrapolation into unsurveyed areas suffers from biases that spatial methods do not resolve; welfare is overestimated in impoverished regions, rural areas, and single sector-dominated economies, whereas it tends to be underestimated in wealthier, urbanized, and diversified economies. Even as spatial models improve overall predictive accuracy, enhancements in traditionally underperforming areas remain marginal. This underscores the need for more representative training datasets and better remotely sensed proxies, especially for poor and rural regions, in future research related to machine learning-based poverty mapping. |
Keywords: | poverty mapping;machine learning;spatial models;East Africa |
JEL: | C21 C55 I32 |
Date: | 2025–09–05 |
URL: | https://d.repec.org/n?u=RePEc:ris:adbewp:021518 |
By: | Marcelo C. Medeiros; Jeronymo M. Pinro |
Abstract: | The forecasting combination puzzle is a well-known phenomenon in forecasting literature, stressing the challenge of outperforming the simple average when aggregating forecasts from diverse methods. This study proposes a Reinforcement Learning - based framework as a dynamic model selection approach to address this puzzle. Our framework is evaluated through extensive forecasting exercises using simulated and real data. Specifically, we analyze the M4 Competition dataset and the Survey of Professional Forecasters (SPF). This research introduces an adaptable methodology for selecting and combining forecasts under uncertainty, offering a promising advancement in resolving the forecasting combination puzzle. |
Date: | 2025–08 |
URL: | https://d.repec.org/n?u=RePEc:arx:papers:2508.20795 |
By: | Ziyao Wang; Svetlozar T Rachev |
Abstract: | Financial returns are known to exhibit heavy tails, volatility clustering and abrupt jumps that are poorly captured by classical diffusion models. Advances in machine learning have enabled highly flexible functional forms for conditional means and volatilities, yet few models deliver interpretable state--dependent tail risk, capture multiple forecast horizons and yield distributions amenable to backtesting and execution. This paper proposes a neural L\'evy jump--diffusion framework that jointly learns, as functions of observable state variables, the conditional drift, diffusion, jump intensity and jump size distribution. We show how a single shared encoder yields multiple forecasting heads corresponding to distinct horizons (daily, weekly, etc.), facilitating multi--horizon density forecasts and risk measures. The state vector includes conventional price and volume features as well as novel complexity measures such as permutation entropy and recurrence quantification analysis determinism, which quantify predictability in the underlying process. Estimation is based on a quasi--maximum likelihood approach that separates diffusion and jump contributions via bipower variation weights and incorporates monotonicity and smoothness regularisation to ensure identifiability. A cost--aware portfolio optimiser translates the model's conditional densities into implementable trading strategies under leverage, turnover and no--trade--band constraints. Extensive empirical analyses on cross--sectional equity data demonstrate improved calibration, sharper tail control and economically significant risk reduction relative to baseline diffusive and GARCH benchmarks. The proposed framework is therefore an interpretable, testable and practically deployable method for state--dependent risk and density forecasting. |
Date: | 2025–08 |
URL: | https://d.repec.org/n?u=RePEc:arx:papers:2509.01041 |
By: | Gabriele Agliardi; Dimitris Alevras; Vaibhaw Kumar; Roberto Lo Nardo; Gabriele Compostella; Sumit Kumar; Manuel Proissl; Bimal Mehta |
Abstract: | The efficient and effective construction of portfolios that adhere to real-world constraints is a challenging optimization task in finance. We investigate a concrete representation of the problem with a focus on design proposals of an Exchange Traded Fund. We evaluate the sampling-based CVaR Variational Quantum Algorithm (VQA), combined with a local-search post-processing, for solving problem instances that beyond a certain size become classically hard. We also propose a problem formulation that is suited for sampling-based VQA. Our utility-scale experiments on IBM Heron processors involve 109 qubits and up to 4200 gates, achieving a relative solution error of 0.49%. Results indicate that a combined quantum-classical workflow achieves better accuracy compared to purely classical local search, and that hard-to-simulate quantum circuits may lead to better convergence than simpler circuits. Our work paves the path to further explore portfolio construction with quantum computers. |
Date: | 2025–08 |
URL: | https://d.repec.org/n?u=RePEc:arx:papers:2508.13557 |
By: | Ziyi Wang; Carmine Ventre; Maria Polukarov |
Abstract: | We advance market-making strategies by integrating Adversarial Reinforcement Learning (ARL), Hawkes Processes, and variable volatility levels while also expanding the action space available to market makers (MMs). To enhance the adaptability and robustness of these strategies -- which can quote always, quote only on one side of the market or not quote at all -- we shift from the commonly used Poisson process to the Hawkes process, which better captures real market dynamics and self-exciting behaviors. We then train and evaluate strategies under volatility levels of 2 and 200. Our findings show that the 4-action MM trained in a low-volatility environment effectively adapts to high-volatility conditions, maintaining stable performance and providing two-sided quotes at least 92\% of the time. This indicates that incorporating flexible quoting mechanisms and realistic market simulations significantly enhances the effectiveness of market-making strategies. |
Date: | 2025–08 |
URL: | https://d.repec.org/n?u=RePEc:arx:papers:2508.16589 |
By: | Gorjian, Mahshid |
Abstract: | This study brings together current advances in the statistical and methodological foundations of spatial economics, focusing on the use of quantitative models and empirical approaches to investigate the distribution of economic activity over geographic space. We combine classical principles with modern approaches that emphasize causal identification, structural estimation, and the use of statistical and computational tools such as spatial econometrics, machine learning, and big data analytics. The study focuses on methodological challenges in spatial data analysis, such as spatial autocorrelation, high dimensionality, and the use of Geographic Information Systems (GIS), while also discussing advances in the design and estimation of quantitative spatial models. The focus is on contemporary empirical applications that use natural experiments, quasi-experimental approaches, and advanced econometric tools to examine the effects of agglomeration, market access, and infrastructure policy. Despite significant advances, significant challenges remain in resilient model identification, dynamic analysis, and the integration of statistical approaches with new types of geographic data. This page focuses on statistical methodologies and serves as a resource for economists and the broader statistics community interested in spatial modeling, causal inference, and policy evaluation. |
Keywords: | statistical methodology, causal inference, spatial econometrics, machine learning, quantitative models, spatial statistics, GIS. |
JEL: | C01 C1 |
Date: | 2025 |
URL: | https://d.repec.org/n?u=RePEc:pra:mprapa:125636 |
By: | Yanlong Wang; Jian Xu; Fei Ma; Hongkang Zhang; Hang Yu; Tiantian Gao; Yu Wang; Haochen You; Shao-Lun Huang; Danny Dongning Sun; Xiao-Ping Zhang |
Abstract: | Financial time series forecasting is both highly significant and challenging. Previous approaches typically standardized time series data before feeding it into forecasting models, but this encoding process inherently leads to a loss of important information. Moreover, past time series models generally require fixed numbers of variables or lookback window lengths, which further limits the scalability of time series forecasting. Besides, the interpretability and the uncertainty in forecasting remain areas requiring further research, as these factors directly impact the reliability and practical value of predictions. To address these issues, we first construct a diverse financial image-text dataset (FVLDB) and develop the Uncertainty-adjusted Group Relative Policy Optimization (UARPO) method to enable the model not only output predictions but also analyze the uncertainty of those predictions. We then proposed FinZero, a multimodal pre-trained model finetuned by UARPO to perform reasoning, prediction, and analytical understanding on the FVLDB financial time series. Extensive experiments validate that FinZero exhibits strong adaptability and scalability. After fine-tuning with UARPO, FinZero achieves an approximate 13.48\% improvement in prediction accuracy over GPT-4o in the high-confidence group, demonstrating the effectiveness of reinforcement learning fine-tuning in multimodal large model, including in financial time series forecasting tasks. |
Date: | 2025–09 |
URL: | https://d.repec.org/n?u=RePEc:arx:papers:2509.08742 |
By: | Yushi Lin; Peng Yang |
Abstract: | Financial markets are critical to global economic stability, yet trade-based manipulation (TBM) often undermines their fairness. Spoofing, a particularly deceptive TBM strategy, exhibits multilevel anomaly patterns that have not been adequately modeled. These patterns are usually concealed within the rich, hierarchical information of the Limit Order Book (LOB), which is challenging to leverage due to high dimensionality and noise. To address this, we propose a representation learning framework combining a cascaded LOB representation pipeline with supervised contrastive learning. Extensive experiments demonstrate that our framework consistently improves detection performance across diverse models, with Transformer-based architectures achieving state-of-the-art results. In addition, we conduct systematic analyses and ablation studies to investigate multilevel anomalies and the contributions of key components, offering broader insights into representation learning and anomaly detection for complex sequential data. Our code will be released later at this URL. |
Date: | 2025–08 |
URL: | https://d.repec.org/n?u=RePEc:arx:papers:2508.17086 |
By: | Shaofeng Kang; Zeying Tian |
Abstract: | We propose a two-level, learning-based portfolio method (RL-BHRP) that spreads risk across sectors and stocks, and adjusts exposures as market conditions change. Using U.S. Equities from 2012 to mid-2025, we design the model using 2012 to 2019 data, and evaluate it out-of-sample from 2020 to 2025 against a sector index built from exchange-traded funds and a static risk-balanced portfolio. Over the test window, the adaptive portfolio compounds wealth by approximately 120 percent, compared with 101 percent for the static comparator and 91 percent for the sector benchmark. The average annual growth is roughly 15 percent, compared to 13 percent and 12 percent, respectively. Gains are achieved without significant deviations from the benchmark and with peak-to-trough losses comparable to those of the alternatives, indicating that the method adds value while remaining diversified and investable. Weight charts show gradual shifts rather than abrupt swings, reflecting disciplined rebalancing and the cost-aware design. Overall, the results support risk-balanced, adaptive allocation as a practical approach to achieving stronger and more stable long-term performance. |
Date: | 2025–08 |
URL: | https://d.repec.org/n?u=RePEc:arx:papers:2508.11856 |
By: | Sebastien Gallet; Julja Prodani |
Abstract: | This paper builds on existing literature on federated learning to introduce an innovative framework, which we call federated modelling. Federated modelling enables collaborative modelling by a group of participants while bypassing the need for disclosing participants’ underlying private data, which are restricted due to legal or institutional requirements. While the uses of this framework can be numerous, the paper presents a proof of concept for a system-wide, granular financial stress test that enables effective cooperation among central banks without the need to disclose the underlying private data and models of the participating central banks or their reporting entities (banks and insurers). Our findings confirm that by leveraging machine learning techniques and using readily available computational tools, the framework allows participants to contribute to the development of shared models whose results are comparable to those using full granular data centralization. This has profound implications for regulatory cooperation and financial stability monitoring across jurisdictions. |
Date: | 2025–08 |
URL: | https://d.repec.org/n?u=RePEc:dnb:dnbocs:2503 |
By: | Jose Blanchet; Jiayi Cheng; Hao Liu; Yang Liu |
Abstract: | We consider a Bayesian diffusion control problem of expected terminal utility maximization. The controller imposes a prior distribution on the unknown drift of an underlying diffusion. The Bayesian optimal control, tracking the posterior distribution of the unknown drift, can be characterized explicitly. However, in practice, the prior will generally be incorrectly specified, and the degree of model misspecification can have a significant impact on policy performance. To mitigate this and reduce overpessimism, we introduce a distributionally robust Bayesian control (DRBC) formulation in which the controller plays a game against an adversary who selects a prior in divergence neighborhood of a baseline prior. The adversarial approach has been studied in economics and efficient algorithms have been proposed in static optimization settings. We develop a strong duality result for our DRBC formulation. Combining these results together with tools from stochastic analysis, we are able to derive a loss that can be efficiently trained (as we demonstrate in our numerical experiments) using a suitable neural network architecture. As a result, we obtain an effective algorithm for computing the DRBC optimal strategy. The methodology for computing the DRBC optimal strategy is greatly simplified, as we show, in the important case in which the adversary chooses a prior from a Kullback-Leibler distributional uncertainty set. |
Date: | 2025–06 |
URL: | https://d.repec.org/n?u=RePEc:arx:papers:2506.19294 |
By: | Jaime Vera-Jaramillo |
Abstract: | This study proposes a unified multi-stage framework to reconstruct consistent monthly and annual labor indicators for all 33 Colombian departments from 1993 to 2025. The approach integrates temporal disaggregation, time-series splicing and interpolation, statistical learning, and institutional covariates to estimate seven key variables: employment, unemployment, labor force participation (PEA), inactivity, working-age population (PET), total population, and informality rate, including in regions without direct survey coverage. The framework enforces labor accounting identities, scales results to demographic projections, and aligns all estimates with national benchmarks to ensure internal coherence. Validation against official departmental GEIH aggregates and city-level informality data for the 23 metropolitan areas yields in-sample Mean Absolute Percentage Errors (MAPEs) below 2.3% across indicators, confirming strong predictive performance. To our knowledge, this is the first dataset to provide spatially exhaustive and temporally consistent monthly labor measures for Colombia. By incorporating both quantitative and qualitative dimensions of employment, the panel enhances the empirical foundation for analysing long-term labor market dynamics, identifying regional disparities, and designing targeted policy interventions. |
Date: | 2025–08 |
URL: | https://d.repec.org/n?u=RePEc:arx:papers:2508.12514 |
By: | Matej Steinbacher; Mitja Steinbacher; Matjaz Steinbacher |
Abstract: | This study examines the impact of different computing implementations of clearing mechanisms on multi-asset price dynamics within an artificial stock market framework. We show that sequential processing of order books introduces a systematic and significant bias by affecting the allocation of traders' capital within a single time step. This occurs because applying budget constraints sequentially grants assets processed earlier preferential access to funds, distorting individual asset demand and consequently their price trajectories. The findings highlight that while the overall price level is primarily driven by macro factors like the money-to-stock ratio, the market's microstructural clearing mechanism plays a critical role in the allocation of value among individual assets. This underscores the necessity for careful consideration and validation of clearing mechanisms in artificial markets to accurately model complex financial behaviors. |
Date: | 2025–09 |
URL: | https://d.repec.org/n?u=RePEc:arx:papers:2509.01683 |
By: | Lee, David |
Abstract: | Financial market data are known to be far from normal and replete with outliers, i.e., “dirty” data that contain errors. Data errors introduce extreme or aberrant data points that can significantly distort parameter estimation results. This paper proposes a robust estimation approach to achieve stable and accurate results. The robust estimation approach is particularly applicable for financial data that often features the three situations we are protecting against: occasional rogue values (outliers), small errors and underlying non-normality. |
Keywords: | robust parameter estimation, financial market data, market data simulation, risk factor. |
JEL: | C13 C15 C53 C63 G17 |
Date: | 2025–08 |
URL: | https://d.repec.org/n?u=RePEc:pra:mprapa:125703 |
By: | Pablo de la Vega |
Abstract: | We analyze the potential economic impacts in Argentina of the European Union Deforestation Regulation (EUDR), which as of January 2026 will prohibit the export to the European Union of certain raw materials and related products if they involve the use of deforested land. We estimate that the EUDR would cover around 6 billion US dollars in exported value, but only 2.84% is not compliant with the EUDR, with soy and cattle being the most affected production chains. We use a dynamic computable general equilibrium model to simulate the impact of the EUDR on the Argentine economy. If the non-compliant production cannot enter the EU market because of the EUDR, the results of the simulations suggest that the potential macroeconomic impacts are limited: GDP would be reduced by an average of 0.14% with respect to the baseline scenario. However, the potential environmental impact is greater. Deforested hectares would be reduced by 2.45% and GHG emissions by 0.19%. Notwithstanding, EUDR due diligence costs may still prevent compliant production from entering the EU market, so the total impacts could be higher. |
Date: | 2025–08 |
URL: | https://d.repec.org/n?u=RePEc:arx:papers:2508.11796 |
By: | Luis Gruber; Gregor Kastner; Anirban Bhattacharya; Debdeep Pati; Natesh Pillai; David Dunson |
Abstract: | Bhattacharya et al. (2015, Journal of the American Statistical Association 110(512): 1479-1490) introduce a novel prior, the Dirichlet-Laplace (DL) prior, and propose a Markov chain Monte Carlo (MCMC) method to simulate posterior draws under this prior in a conditionally Gaussian setting. The original algorithm samples from conditional distributions in the wrong order, i.e., it does not correctly sample from the joint posterior distribution of all latent variables. This note details the issue and provides two simple solutions: A correction to the original algorithm and a new algorithm based on an alternative, yet equivalent, formulation of the prior. This corrigendum does not affect the theoretical results in Bhattacharya et al. (2015). |
Date: | 2025–08 |
URL: | https://d.repec.org/n?u=RePEc:arx:papers:2508.11982 |
By: | Kerstin Bernoth |
Abstract: | This paper investigates the effectiveness of the European Central Bank’s (ECB) communication in shaping market expectations and real economic outcomes. Using a transformer-based large language model (LLM) fine-tuned to ECB communication, the tone of monetary policy statements from 2003 to 2025 is classified, constructing a novel ECB Communication Stance Indicator. This indicator contains forward-looking information beyond standard macro-financial variables. Identified communication shocks are distinct from monetary policy and central bank information shocks. A structural Bayesian VAR reveals that hawkish communication signals favorable economic prospects, raising output, equity prices, and inflation, but also increases bond market stress. These findings highlight communication as an independent and effective tool of monetary policy, while also underscoring the importance of carefully calibrating tone to balance market expectations, and financial stability. |
Keywords: | Monetary Policy, Central Bank Communication, Text Sentiment, Transformerbased Large Language Model, Bayesian Vector Autoregression, Local Projections |
JEL: | C32 E43 E47 E52 E58 |
Date: | 2025 |
URL: | https://d.repec.org/n?u=RePEc:diw:diwwpp:dp2137 |
By: | Adrian Ifrim; Robert Kollmann; Philipp Pfeiffer; Marco Ratto; Werner Roeger |
Abstract: | Based on an estimated two-region dynamic general equilibrium model, we show that the persistent productivity growth differential between the Euro Area (EA) and rest of the world (RoW) has been a key driver of the EA trade surplus since the launch of the Euro. A secular decline in the EA's spending home bias and a trend decrease in relative EA import prices account for the stability of the EA real exchange rate, despite slower EA output growth. By incorporating trend shocks to growth and trade, the analysis departs from much of the open-economy macroeconomics literature which has focused on stationary disturbances. Our results highlight the relevance of non-stationary shocks for the analysis of external adjustment. |
Keywords: | global growth divergences, trade balance, real exchange rate, estimated DSGE model, Euro Area, demand and supply shocks, persistent growth shocks |
JEL: | F4 F3 E2 E3 C5 |
Date: | 2025–09 |
URL: | https://d.repec.org/n?u=RePEc:een:camaaa:2025-50 |