nep-cmp New Economics Papers
on Computational Economics
Issue of 2025–12–15
34 papers chosen by
Stan Miles, Thompson Rivers University


  1. Generative economic modeling By Hanno Kase; Matthias Rottner; Fabio Stohler
  2. Cryptocurrency Portfolio Management with Reinforcement Learning: Soft Actor--Critic and Deep Deterministic Policy Gradient Algorithms By Kamal Paykan
  3. Modelling the Doughnut of social and planetary boundaries with frugal machine learning By Stefano Vrizzi; Daniel W. O'Neill
  4. Solving Heterogeneous Agent Models with Physics-informed Neural Networks By Marta Grzeskiewicz
  5. GDP Nowcasting Performance of Traditional Econometric Models vs Machine-Learning Algorithms: Simulation and Case Studies By Klakow Akepanidtaworn; Korkrid Akepanidtaworn
  6. Using Machine Learning Method to Estimate the Heterogeneous Impacts of the Updated Nutrition Facts Panel By Zhang, Yuxiang; Liu, Yizao; Sears, James M.
  7. Continuous-time reinforcement learning for optimal switching over multiple regimes By Yijie Huang; Mengge Li; Xiang Yu; Zhou Zhou
  8. Explainable Machine Learning for Macroeconomic and Financial Nowcasting: A Decision-Grade Framework for Business and Policy By Luca Attolico
  9. Investigating Factors Influencing Dietary Quality in China: Machine Learning Approaches By Feng, Yuan; Liu, Shuang; Zhang, Man; Jin, Yanhong; Yu, Xiaohua
  10. Arbitrage-Free Bond and Yield Curve Forecasting with Neural Filters under HJM Constraints By Xiang Gao; Cody Hyndman
  11. DeXposure: A Dataset and Benchmarks for Inter-protocol Credit Exposure in Decentralized Financial Networks By Wenbin Wu; Kejiang Qian; Alexis Lui; Christopher Jack; Yue Wu; Peter McBurney; Fengxiang He; Bryan Zhang
  12. Financial Text Classification Based On rLoRA Finetuning On Qwen3-8B model By Zhiming Lian
  13. A Hybrid Architecture for Options Wheel Strategy Decisions: LLM-Generated Bayesian Networks for Transparent Trading By Xiaoting Kuang; Boken Lin
  14. Enhancing the Efficiency of National R&D Programs Using Machine Learning-Based Anomaly Detection By Sang-Kyu Lee
  15. Hybrid LSTM and PPO Networks for Dynamic Portfolio Optimization By Jun Kevin; Pujianto Yugopuspito
  16. Predicting Price Movements in High-Frequency Financial Data with Spiking Neural Networks By Brian Ezinwoke; Oliver Rhodes
  17. Tacit Bidder-Side Collusion: Artificial Intelligence in Dynamic Auctions By Sriram Tolety
  18. Workflow is All You Need: Escaping the "Statistical Smoothing Trap" via High-Entropy Information Foraging and Adversarial Pacing By Zhongjie Jiang
  19. Optimizing Information Asset Investment Strategies in the Exploratory Phase of the Oil and Gas Industry: A Reinforcement Learning Approach By Paulo Roberto de Melo Barros Junior; Monica Alexandra Vilar Ribeiro De Meireles; Jose Luis Lima de Jesus Silva
  20. Standard Occupation Classifier -- A Natural Language Processing Approach By Sidharth Rony; Jack Patman
  21. Re(Visiting) Time Series Foundation Models in Finance By Eghbal Rahimikia; Hao Ni; Weiguan Wang
  22. Detecting AI Hallucinations in Finance: An Information-Theoretic Method Cuts Hallucination Rate by 92% By Mainak Singha
  23. Stochastic Dominance Constrained Optimization with S-shaped Utilities: Poor-Performance-Region Algorithm and Neural Network By Zeyun Hu; Yang Liu
  24. Can technology augment order writing capacity at regulators? By Natasha Aggarwal; Satyavrat Bondre; Amrutha Desikan; Bhavin Patel; Dipyaman Sanyal
  25. Narratives to Numbers: Large Language Models and Economic Policy Uncertainty By Ethan Hartley
  26. Semantic Faithfulness and Entropy Production Measures to Tame Your LLM Demons and Manage Hallucinations By Igor Halperin
  27. Sentiment Analysis of Financial Text Using Quantum Language Processing QDisCoCirc By Takayuki Sakuma
  28. Data Science and Artificial Intelligence for Statistics Education: Creating Smart Future of Teaching and Learning By Popoola, Osuolale Peter; Kumafan, Dzaan
  29. Adaptive agent-based modeling in finance : selected applications By Marcello Esposito
  30. Statistical Arbitrage in Polish Equities Market Using Deep Learning Techniques By Marek Adamczyk; Micha{\l} D\k{a}browski
  31. The Necessity of Imperfection:Reversing Model Collapse via Simulating Cognitive Boundedness By Zhongjie Jiang
  32. Computing Evolutionarily Stable Strategies in Multiplayer Games By Sam Ganzfried
  33. Random processes for long-term market simulations By Gilles Zumbach
  34. COMPLEXITY: A Generalized Approach in Stata for Specialization Complexity Indices By Charlie Joyez

  1. By: Hanno Kase; Matthias Rottner; Fabio Stohler
    Abstract: We introduce a novel approach for solving quantitative economic models: generative economic modeling. Our method combines neural networks with conventional solution techniques. Specifically, we train neural networks on simplified versions of the economic model to approximate the complete model's dynamic behavior. Relying on these less complex submodels circumvents the curse of dimensionality, allowing the use of well-established numerical methods. We demonstrate our approach across settings with analytical characterizations, nonlinear dynamics, and heterogeneous agents, employing asset pricing and business cycle models. Finally, we solve a high-dimensional HANK model with an occasionally binding financial friction to highlight how aggregate risk amplifies the precautionary motive.
    Keywords: machine learning, neural networks, nonlinearities, heterogeneous agents
    JEL: C11 C45 D31 E32 E52
    Date: 2025–12
    URL: https://d.repec.org/n?u=RePEc:bis:biswps:1312
  2. By: Kamal Paykan (Department of Mathematics, Tafresh University, Tafresh, Iran)
    Abstract: This paper proposes a reinforcement learning--based framework for cryptocurrency portfolio management using the Soft Actor--Critic (SAC) and Deep Deterministic Policy Gradient (DDPG) algorithms. Traditional portfolio optimization methods often struggle to adapt to the highly volatile and nonlinear dynamics of cryptocurrency markets. To address this, we design an agent that learns continuous trading actions directly from historical market data through interaction with a simulated trading environment. The agent optimizes portfolio weights to maximize cumulative returns while minimizing downside risk and transaction costs. Experimental evaluations on multiple cryptocurrencies demonstrate that the SAC and DDPG agents outperform baseline strategies such as equal-weighted and mean--variance portfolios. The SAC algorithm, with its entropy-regularized objective, shows greater stability and robustness in noisy market conditions compared to DDPG. These results highlight the potential of deep reinforcement learning for adaptive and data-driven portfolio management in cryptocurrency markets.
    Date: 2025–11
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2511.20678
  3. By: Stefano Vrizzi; Daniel W. O'Neill
    Abstract: The 'Doughnut' of social and planetary boundaries has emerged as a popular framework for assessing environmental and social sustainability. Here, we provide a proof-of-concept analysis that shows how machine learning (ML) methods can be applied to a simple macroeconomic model of the Doughnut. First, we show how ML methods can be used to find policy parameters that are consistent with 'living within the Doughnut'. Second, we show how a reinforcement learning agent can identify the optimal trajectory towards desired policies in the parameter space. The approaches we test, which include a Random Forest Classifier and $Q$-learning, are frugal ML methods that are able to find policy parameter combinations that achieve both environmental and social sustainability. The next step is the application of these methods to a more complex ecological macroeconomic model.
    Date: 2025–12
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2512.02200
  4. By: Marta Grzeskiewicz
    Abstract: Understanding household behaviour is essential for modelling macroeconomic dynamics and designing effective policy. While heterogeneous agent models offer a more realistic alternative to representative agent frameworks, their implementation poses significant computational challenges, particularly in continuous time. The Aiyagari-Bewley-Huggett (ABH) framework, recast as a system of partial differential equations, typically relies on grid-based solvers that suffer from the curse of dimensionality, high computational cost, and numerical inaccuracies. This paper introduces the ABH-PINN solver, an approach based on Physics-Informed Neural Networks (PINNs), which embeds the Hamilton-Jacobi-Bellman and Kolmogorov Forward equations directly into the neural network training objective. By replacing grid-based approximation with mesh-free, differentiable function learning, the ABH-PINN solver benefits from the advantages of PINNs of improved scalability, smoother solutions, and computational efficiency. Preliminary results show that the PINN-based approach is able to obtain economically valid results matching the established finite-difference solvers.
    Date: 2025–11
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2511.20283
  5. By: Klakow Akepanidtaworn; Korkrid Akepanidtaworn
    Abstract: Are Machine Learning (ML) algorithms superior to traditional econometric models for GDP nowcasting in a time series setting? Based on our evaluation of all models from both classes ever used in nowcasting across simulation and six country cases, traditional econometric models tend to outperform ML algorithms. Among the ML algorithms, linear ML algorithm – Lasso and Elastic Net – perform best in nowcasting, even surpassing traditional econometric models in cases of long GDP data and rich high-frequency indicators. Among the traditional econometric models, the Bridge and Dynamic Factor deliver the strongest empirical results, while Three-Pass Regression Filter performs well in our simulation. Due to the relatively short length of GDP series, complex and non-linear ML algorithms are prone to overfitting, which compromises their out-of-sample performance.
    Keywords: Nowcasting; Machine Learning; Forecast evaluation; Real-time data
    Date: 2025–12–05
    URL: https://d.repec.org/n?u=RePEc:imf:imfwpa:2025/252
  6. By: Zhang, Yuxiang; Liu, Yizao; Sears, James M.
    Keywords: Food Consumption/Nutrition/Food Safety, Health Economics and Policy, Consumer/Household Economics
    Date: 2024
    URL: https://d.repec.org/n?u=RePEc:ags:aaea24:343727
  7. By: Yijie Huang; Mengge Li; Xiang Yu; Zhou Zhou
    Abstract: This paper studies the continuous-time reinforcement learning (RL) for optimal switching problems across multiple regimes. We consider a type of exploratory formulation under entropy regularization where the agent randomizes both the timing of switches and the selection of regimes through the generator matrix of an associated continuous-time finite-state Markov chain. We establish the well-posedness of the associated system of Hamilton-Jacobi-Bellman (HJB) equations and provide a characterization of the optimal policy. The policy improvement and the convergence of the policy iterations are rigorously established by analyzing the system of equations. We also show the convergence of the value function in the exploratory formulation towards the value function in the classical formulation as the temperature parameter vanishes. Finally, a reinforcement learning algorithm is devised and implemented by invoking the policy evaluation based on the martingale characterization. Our numerical examples with the aid of neural networks illustrate the effectiveness of the proposed RL algorithm.
    Date: 2025–12
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2512.04697
  8. By: Luca Attolico
    Abstract: Macroeconomic nowcasting sits at the intersection of traditional econometrics, data-rich information systems, and AI applications in business, economics, and policy. Machine learning (ML) methods are increasingly used to nowcast quarterly GDP growth, but adoption in high-stakes settings requires that predictive accuracy be matched by interpretability and robust uncertainty quantification. This article reviews recent developments in macroeconomic nowcasting and compares econometric benchmarks with ML approaches in data-rich and shock-prone environments, emphasizing the use of nowcasts as decision inputs rather than as mere error-minimization exercises. The discussion is organized along three axes. First, we contrast penalized regressions, dimension-reduction techniques, tree ensembles, and neural networks with autoregressive models, Dynamic Factor Models, and Random Walks, emphasizing how each family handles small samples, collinearity, mixed frequencies, and regime shifts. Second, we examine explainability tools (intrinsic measures and model-agnostic XAI methods), focusing on temporal stability, sign coherence, and their ability to sustain credible economic narratives and nowcast revisions. Third, we analyze non-parametric uncertainty quantification via block bootstrapping for predictive intervals and confidence bands on feature importance under serial dependence and ragged edge. We translate these elements into a reference workflow for "decision-grade" nowcasting systems, including vintage management, time-aware validation, and automated reliability audits, and we outline a research agenda on regime-dependent model comparison, bootstrap design for latent components, and temporal stability of explanations. Explainable ML and uncertainty quantification emerge as structural components of a responsible forecasting pipeline, not optional refinements.
    Date: 2025–11
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2512.00399
  9. By: Feng, Yuan; Liu, Shuang; Zhang, Man; Jin, Yanhong; Yu, Xiaohua
    Keywords: Food Consumption/Nutrition/Food Safety
    Date: 2024
    URL: https://d.repec.org/n?u=RePEc:ags:aaea24:343836
  10. By: Xiang Gao; Cody Hyndman
    Abstract: We develop an arbitrage-free deep learning framework for yield curve and bond price forecasting based on the Heath-Jarrow-Morton (HJM) term-structure model and a dynamic Nelson-Siegel parameterization of forward rates. Our approach embeds a no-arbitrage drift restriction into a neural state-space architecture by combining Kalman, extended Kalman, and particle filters with recurrent neural networks (LSTM/CLSTM), and introduces an explicit arbitrage error regularization (AER) term during training. The model is applied to U.S. Treasury and corporate bond data, and its performance is evaluated for both yield-space and price-space predictions at 1-day and 5-day horizons. Empirically, arbitrage regularization leads to its strongest improvements at short maturities, particularly in 5-day-ahead forecasts, increasing market-consistency as measured by bid-ask hit rates and reducing dollar-denominated prediction errors.
    Date: 2025–11
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2511.17892
  11. By: Wenbin Wu; Kejiang Qian; Alexis Lui; Christopher Jack; Yue Wu; Peter McBurney; Fengxiang He; Bryan Zhang
    Abstract: We curate the DeXposure dataset, the first large-scale dataset for inter-protocol credit exposure in decentralized financial networks, covering global markets of 43.7 million entries across 4.3 thousand protocols, 602 blockchains, and 24.3 thousand tokens, from 2020 to 2025. A new measure, value-linked credit exposure between protocols, is defined as the inferred financial dependency relationships derived from changes in Total Value Locked (TVL). We develop a token-to-protocol model using DefiLlama metadata to infer inter-protocol credit exposure from the token's stock dynamics, as reported by the protocols. Based on the curated dataset, we develop three benchmarks for machine learning research with financial applications: (1) graph clustering for global network measurement, tracking the structural evolution of credit exposure networks, (2) vector autoregression for sector-level credit exposure dynamics during major shocks (Terra and FTX), and (3) temporal graph neural networks for dynamic link prediction on temporal graphs. From the analysis, we observe (1) a rapid growth of network volume, (2) a trend of concentration to key protocols, (3) a decline of network density (the ratio of actual connections to possible connections), and (4) distinct shock propagation across sectors, such as lending platforms, trading exchanges, and asset management protocols. The DeXposure dataset and code have been released publicly. We envision they will help with research and practice in machine learning as well as financial risk monitoring, policy analysis, DeFi market modeling, amongst others. The dataset also contributes to machine learning research by offering benchmarks for graph clustering, vector autoregression, and temporal graph analysis.
    Date: 2025–11
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2511.22314
  12. By: Zhiming Lian
    Abstract: Financial text classification has increasingly become an important aspect in quantitative trading systems and related tasks, such as financial sentiment analysis and the classification of financial news. In this paper, we assess the performance of the large language model Qwen3-8B on both tasks. Qwen3-8B is a state-of-the-art model that exhibits strong instruction-following and multilingual capabilities, and is distinct from standard models, primarily because it is specifically optimized for efficient fine tuning and high performance on reasoning-based benchmarks, making it suitable for financial applications. To adapt this model, we apply Noisy Embedding Instruction Finetuning and based on our previous work, this method increases robustness by injecting controlled noise into the embedding layers during supervised adaptation. We improve efficiency further with Rank-stabilized Low-Rank Adaptation low-rank optimization approach, and FlashAttention, which allow for faster training with lower GPU memory. For both tasks, we benchmark Qwen3-8B against standard classical transformer models, such as T5, BERT, and RoBERTa, and large models at scale, such as LLaMA1-7B, LLaMA2-7B, and Baichuan2-7B. The findings reveal that Qwen3-8B consistently surpasses these baselines by obtaining better classification accuracy and needing fewer training epochs. The synergy of instruction-based fine-tuning and memory-efficient optimization methods suggests Qwen3-8B can potentially serve as a scalable, economical option for real-time financial NLP applications. Qwen3-8B provides a very promising base for advancing dynamic quantitative trading systems in the future.
    Date: 2025–11
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2512.00630
  13. By: Xiaoting Kuang; Boken Lin
    Abstract: Large Language Models (LLMs) excel at understanding context and qualitative nuances but struggle with the rigorous and transparent reasoning required in high-stakes quantitative domains such as financial trading. We propose a model-first hybrid architecture for the options "wheel" strategy that combines the strengths of LLMs with the robustness of a Bayesian Network. Rather than using the LLM as a black-box decision-maker, we employ it as an intelligent model builder. For each trade decision, the LLM constructs a context-specific Bayesian network by interpreting current market conditions, including prices, volatility, trends, and news, and hypothesizing relationships among key variables. The LLM also selects relevant historical data from an 18.75-year, 8, 919-trade dataset to populate the network's conditional probability tables. This selection focuses on scenarios analogous to the present context. The instantiated Bayesian network then performs transparent probabilistic inference, producing explicit probability distributions and risk metrics to support decision-making. A feedback loop enables the LLM to analyze trade outcomes and iteratively refine subsequent network structures and data selection, learning from both successes and failures. Empirically, our hybrid system demonstrates effective performance on the wheel strategy. Over nearly 19 years of out-of-sample testing, it achieves a 15.3% annualized return with significantly superior risk-adjusted performance (Sharpe ratio 1.08 versus 0.62 for market benchmarks) and dramatically lower drawdown (-8.2% versus -60%) while maintaining a 0% assignment rate through strategic option rolling. Crucially, each trade decision is fully explainable, involving on average 27 recorded decision factors (e.g., volatility level, option premium, risk indicators, market context).
    Date: 2025–11
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2512.01123
  14. By: Sang-Kyu Lee (Korea Institute for Industrial Economics and Trade)
    Abstract: This study is grounded on the premise that, given the transformative advances in artificial intelligence (AI) technologies occurring across the industrial landscape, AI tools should be actively implemented into the design and implementation of industrial policy. We argue that this is especially true for R&D policy, which is central to national competitiveness in science and technology, and which must consider multiple diverse variables, including the global economy, the overall industrial environment, corporate management, and technological capabilities.<p> For this study, I apply machine learning (ML)-based anomaly detection (AD) to analyze high-performing national R&D projects, and specifically assess ML-based AD that considers both input and output variables and analyzes structural patterns. Building on these analytical results, I propose firm-size-specific differentiated policy measures designed to enhance R&D performance.<p> The goal of this study is to establish a policy-decision framework that improves timeliness and precision in the operation and management of national R&D programs and, in the longer term, contributes to the realization of AI-based policy planning and operational management.
    Keywords: machine learning; artificial intelligence; AI; anomaly detection; DEA; SHAP; research and development; R&D; government R&D; industrial policy; South Korea
    JEL: I23 I28 O32 O38
    Date: 2025–10–31
    URL: https://d.repec.org/n?u=RePEc:ris:kieter:021804
  15. By: Jun Kevin; Pujianto Yugopuspito
    Abstract: This paper introduces a hybrid framework for portfolio optimization that fuses Long Short-Term Memory (LSTM) forecasting with a Proximal Policy Optimization (PPO) reinforcement learning strategy. The proposed system leverages the predictive power of deep recurrent networks to capture temporal dependencies, while the PPO agent adaptively refines portfolio allocations in continuous action spaces, allowing the system to anticipate trends while adjusting dynamically to market shifts. Using multi-asset datasets covering U.S. and Indonesian equities, U.S. Treasuries, and major cryptocurrencies from January 2018 to December 2024, the model is evaluated against several baselines, including equal-weight, index-style, and single-model variants (LSTM-only and PPO-only). The framework's performance is benchmarked against equal-weighted, index-based, and single-model approaches (LSTM-only and PPO-only) using annualized return, volatility, Sharpe ratio, and maximum drawdown metrics, each adjusted for transaction costs. The results indicate that the hybrid architecture delivers higher returns and stronger resilience under non-stationary market regimes, suggesting its promise as a robust, AI-driven framework for dynamic portfolio optimization.
    Date: 2025–11
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2511.17963
  16. By: Brian Ezinwoke; Oliver Rhodes
    Abstract: Modern high-frequency trading (HFT) environments are characterized by sudden price spikes that present both risk and opportunity, but conventional financial models often fail to capture the required fine temporal structure. Spiking Neural Networks (SNNs) offer a biologically inspired framework well-suited to these challenges due to their natural ability to process discrete events and preserve millisecond-scale timing. This work investigates the application of SNNs to high-frequency price-spike forecasting, enhancing performance via robust hyperparameter tuning with Bayesian Optimization (BO). This work converts high-frequency stock data into spike trains and evaluates three architectures: an established unsupervised STDP-trained SNN, a novel SNN with explicit inhibitory competition, and a supervised backpropagation network. BO was driven by a novel objective, Penalized Spike Accuracy (PSA), designed to ensure a network's predicted price spike rate aligns with the empirical rate of price events. Simulated trading demonstrated that models optimized with PSA consistently outperformed their Spike Accuracy (SA)-tuned counterparts and baselines. Specifically, the extended SNN model with PSA achieved the highest cumulative return (76.8%) in simple backtesting, significantly surpassing the supervised alternative (42.54% return). These results validate the potential of spiking networks, when robustly tuned with task-specific objectives, for effective price spike forecasting in HFT.
    Date: 2025–12
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2512.05868
  17. By: Sriram Tolety
    Abstract: We study whether large language models acting as autonomous bidders can tacitly collude by coordinating when to accept platform posted payouts in repeated Dutch auctions, without any communication. We present a minimal repeated auction model that yields a simple incentive compatibility condition and a closed form threshold for sustainable collusion for subgame-perfect Nash equilibria. In controlled simulations with multiple language models, we observe systematic supra-competitive prices in small auction settings and a return to competitive behavior as the number of bidders in the market increases, consistent with the theoretical model. We also find LLMs use various mechanisms to facilitate tacit coordination, such as focal point acceptance timing versus patient strategies that track the theoretical incentives. The results provide, to our knowledge, the first evidence of bidder side tacit collusion by LLMs and show that market structure levers can be more effective than capability limits for mitigation.
    Date: 2025–11
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2511.21802
  18. By: Zhongjie Jiang
    Abstract: Central to long-form text generation in vertical domains is the "impossible trinity" confronting current large language models (LLMs): the simultaneous achievement of low hallucination, deep logical coherence, and personalized expression. This study establishes that this bottleneck arises from existing generative paradigms succumbing to the Statistical Smoothing Trap, a phenomenon that overlooks the high-entropy information acquisition and structured cognitive processes integral to expert-level writing. To address this limitation, we propose the DeepNews Framework, an agentic workflow that explicitly models the implicit cognitive processes of seasoned financial journalists. The framework integrates three core modules: first, a dual-granularity retrieval mechanism grounded in information foraging theory, which enforces a 10:1 saturated information input ratio to mitigate hallucinatory outputs; second, schema-guided strategic planning, a process leveraging domain expert knowledge bases (narrative schemas) and Atomic Blocks to forge a robust logical skeleton; third, adversarial constraint prompting, a technique deploying tactics including Rhythm Break and Logic Fog to disrupt the probabilistic smoothness inherent in model-generated text. Experiments delineate a salient Knowledge Cliff in deep financial reporting: content truthfulness collapses when retrieved context falls below 15, 000 characters, while a high-redundancy input exceeding 30, 000 characters stabilizes the Hallucination-Free Rate (HFR) above 85%. In an ecological validity blind test conducted with a top-tier Chinese technology media outlet, the DeepNews system--built on a previous-generation model (DeepSeek-V3-0324)-achieved a 25% submission acceptance rate, significantly outperforming the 0% acceptance rate of zero-shot generation by a state-of-the-art (SOTA) model (GPT-5).
    Date: 2025–12
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2512.10121
  19. By: Paulo Roberto de Melo Barros Junior; Monica Alexandra Vilar Ribeiro De Meireles; Jose Luis Lima de Jesus Silva
    Abstract: Our work investigates the economic efficiency of the prevailing "ladder-step" investment strategy in oil and gas exploration, which advocates for the incremental acquisition of geological information throughout the project lifecycle. By employing a multi-agent Deep Reinforcement Learning (DRL) framework, we model an alternative strategy that prioritizes the early acquisition of high-quality information assets. We simulate the entire upstream value chain-comprising competitive bidding, exploration, and development phases-to evaluate the economic impact of this approach relative to traditional methods. Our results demonstrate that front-loading information investment significantly reduces the costs associated with redundant data acquisition and enhances the precision of reserve valuation. Specifically, we find that the alternative strategy outperforms traditional methods in highly competitive environments by mitigating the "winner's curse" through more accurate bidding. Furthermore, the economic benefits are most pronounced during the development phase, where superior data quality minimizes capital misallocation. These findings suggest that optimal investment timing is structurally dependent on market competition rather than solely on price volatility, offering a new paradigm for capital allocation in extractive industries.
    Date: 2025–11
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2512.00243
  20. By: Sidharth Rony; Jack Patman
    Abstract: Standard Occupational Classifiers (SOC) are systems used to categorize and classify different types of jobs and occupations based on their similarities in terms of job duties, skills, and qualifications. Integrating these facets with Big Data from job advertisement offers the prospect to investigate labour demand that is specific to various occupations. This project investigates the use of recent developments in natural language processing to construct a classifier capable of assigning an occupation code to a given job advertisement. We develop various classifiers for both UK ONS SOC and US O*NET SOC, using different Language Models. We find that an ensemble model, which combines Google BERT and a Neural Network classifier while considering job title, description, and skills, achieved the highest prediction accuracy. Specifically, the ensemble model exhibited a classification accuracy of up to 61% for the lower (or fourth) tier of SOC, and 72% for the third tier of SOC. This model could provide up to date, accurate information on the evolution of the labour market using job advertisements.
    Date: 2025–11
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2511.23057
  21. By: Eghbal Rahimikia; Hao Ni; Weiguan Wang
    Abstract: Financial time series forecasting is central to trading, portfolio optimization, and risk management, yet it remains challenging due to noisy, non-stationary, and heterogeneous data. Recent advances in time series foundation models (TSFMs), inspired by large language models, offer a new paradigm for learning generalizable temporal representations from large and diverse datasets. This paper presents the first comprehensive empirical study of TSFMs in global financial markets. Using a large-scale dataset of daily excess returns across diverse markets, we evaluate zero-shot inference, fine-tuning, and pre-training from scratch against strong benchmark models. We find that off-the-shelf pre-trained TSFMs perform poorly in zero-shot and fine-tuning settings, whereas models pre-trained from scratch on financial data achieve substantial forecasting and economic improvements, underscoring the value of domain-specific adaptation. Increasing the dataset size, incorporating synthetic data augmentation, and applying hyperparameter tuning further enhance performance.
    Date: 2025–11
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2511.18578
  22. By: Mainak Singha
    Abstract: Large language models (LLMs) produce fluent but unsupported answers - hallucinations - limiting safe deployment in high-stakes domains. We propose ECLIPSE, a framework that treats hallucination as a mismatch between a model's semantic entropy and the capacity of available evidence. We combine entropy estimation via multi-sample clustering with a novel perplexity decomposition that measures how models use retrieved evidence. We prove that under mild conditions, the resulting entropy-capacity objective is strictly convex with a unique stable optimum. We evaluate on a controlled financial question answering dataset with GPT-3.5-turbo (n=200 balanced samples with synthetic hallucinations), where ECLIPSE achieves ROC AUC of 0.89 and average precision of 0.90, substantially outperforming a semantic entropy-only baseline (AUC 0.50). A controlled ablation with Claude-3-Haiku, which lacks token-level log probabilities, shows AUC dropping to 0.59 with coefficient magnitudes decreasing by 95% - demonstrating that ECLIPSE is a logprob-native mechanism whose effectiveness depends on calibrated token-level uncertainties. The perplexity decomposition features exhibit the largest learned coefficients, confirming that evidence utilization is central to hallucination detection. We position this work as a controlled mechanism study; broader validation across domains and naturally occurring hallucinations remains future work.
    Date: 2025–12
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2512.03107
  23. By: Zeyun Hu; Yang Liu
    Abstract: We investigate the static portfolio selection problem of S-shaped and non-concave utility maximization under first-order and second-order stochastic dominance (SD) constraints. In many S-shaped utility optimization problems, one should require a liquidation boundary to guarantee the existence of a finite concave envelope function. A first-order SD (FSD) constraint can replace this requirement and provide an alternative for risk management. We explicitly solve the optimal solution under a general S-shaped utility function with a first-order stochastic dominance constraint. However, the second-order SD (SSD) constrained problem under non-concave utilities is difficult to solve analytically due to the invalidity of Sion's maxmin theorem. For this sake, we propose a numerical algorithm to obtain a plausible and sub-optimal solution for general non-concave utilities. The key idea is to detect the poor performance region with respect to the SSD constraints, characterize its structure and modify the distribution on that region to obtain (sub-)optimality. A key financial insight is that the decision maker should follow the SD constraint on the poor performance scenario while conducting the unconstrained optimal strategy otherwise. We provide numerical experiments to show that our algorithm effectively finds a sub-optimal solution in many cases. Finally, we develop an algorithm-guided piecewise-neural-network framework to learn the solution of the SSD problem, which demonstrates accelerated convergence compared to standard neural network approaches.
    Date: 2025–11
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2512.00299
  24. By: Natasha Aggarwal (TrustBridge Rule of Law Foundation); Satyavrat Bondre (Dono Consulting); Amrutha Desikan (TrustBridge Rule of Law Foundation); Bhavin Patel (TrustBridge Rule of Law Foundation); Dipyaman Sanyal (Dono Consulting)
    Abstract: This paper critically examines the opportunities and challenges of using technology, in particular Large Language Models (LLMs), to assist regulatory order writing in quasi-judicial settings, with a focus on the Indian context. The paper proposes augmenting rather than replacing human decision-makers, aiming to improve regulatory order writing practice through responsible use of LLMs. It identifies the core principles of administrative law that must be upheld in these settings — such as application of mind, reasoned orders, non arbitrariness, rules against bias, and transparency — and analyses how inherent limitations of LLMs, including their probabilistic reasoning, opacity, potential for bias, confabulation, and lack of metacognition, may undermine these principles. The paper reviews international frameworks and case studies from various jurisdictions, highlighting common design principles like human oversight, transparency, nondiscrimination, and security. It proposes a comprehensive Problem-Solution-Evaluation (PSE) framework for responsibly integrating LLMs into order writing processes. This framework maps specific technical, design, and systemic solutions to each identified risk, and outlines evaluation strategies — end-to-end, component-wise, human-in-theloop, and automated — to ensure ongoing alignment with legal standards. The article concludes with practical recommendations for the development and deployment of LLM-based systems in regulatory environments.
    Date: 2025–12
    URL: https://d.repec.org/n?u=RePEc:bjd:wpaper:16
  25. By: Ethan Hartley
    Abstract: This study evaluates large language models as estimable classifiers and clarifies how modeling choices shape downstream measurement error. Revisiting the Economic Policy Uncertainty index, we show that contemporary classifiers substantially outperform dictionary rules, better track human audit assessments, and extend naturally to noisy historical and multilingual news. We use these tools to construct a new nineteenth-century U.S. index from more than 360 million newspaper articles and exploratory cross-country indices with a single multilingual model. Taken together, our results show that LLMs can systematically improve text-derived measures and should be integrated as explicit measurement tools in empirical economics.
    Date: 2025–11
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2511.17866
  26. By: Igor Halperin
    Abstract: Evaluating faithfulness of Large Language Models (LLMs) to a given task is a complex challenge. We propose two new unsupervised metrics for faithfulness evaluation using insights from information theory and thermodynamics. Our approach treats an LLM as a bipartite information engine where hidden layers act as a Maxwell demon controlling transformations of context $C $ into answer $A$ via prompt $Q$. We model Question-Context-Answer (QCA) triplets as probability distributions over shared topics. Topic transformations from $C$ to $Q$ and $A$ are modeled as transition matrices ${\bf Q}$ and ${\bf A}$ encoding the query goal and actual result, respectively. Our semantic faithfulness (SF) metric quantifies faithfulness for any given QCA triplet by the Kullback-Leibler (KL) divergence between these matrices. Both matrices are inferred simultaneously via convex optimization of this KL divergence, and the final SF metric is obtained by mapping the minimal divergence onto the unit interval [0, 1], where higher scores indicate greater faithfulness. Furthermore, we propose a thermodynamics-based semantic entropy production (SEP) metric in answer generation, and show that high faithfulness generally implies low entropy production. The SF and SEP metrics can be used jointly or separately for LLM evaluation and hallucination control. We demonstrate our framework on LLM summarization of corporate SEC 10-K filings.
    Date: 2025–12
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2512.05156
  27. By: Takayuki Sakuma
    Abstract: We apply quantum distributional compositional circuit (QDisCoCirc) to 3-class sentiment analysis of financial text. In our classical simulations, we keep the Hilbert-space dimension manageable by decomposing each sentence into short contiguous chunks. Each chunk is mapped to a shallow quantum circuit, and the resulting Bloch vectors are used as a sequence of quantum tokens. Simple averaging of chunk vectors ignores word order and syntactic roles. We therefore add a small Transformer encoder over the raw Bloch-vector sequence and attach a CCG-based type embedding to each chunk. This hybrid design preserves physically interpretable semantic axes of quantum tokens while allowing the classical side to model word order and long-range dependencies. The sequence model improves test macro-F1 over the averaging baseline and chunk-level attribution further shows that evidential mass concentrates on a small number of chunks, that type embeddings are used more reliably for correctly predicted sentences. For real-world quantum language processing applications in finance, future key challenges include circuit designs that avoid chunking and the design of inter-chunk fusion layers.
    Date: 2025–11
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2511.18804
  28. By: Popoola, Osuolale Peter; Kumafan, Dzaan
    Abstract: Integrating data science and artificial intelligence(AI) into statistics education have the potential to raise academic standards, improve the overall quality of statistics education. Data science is the "what" and "why" of student performance and learning patterns, and AI is the "how" the intelligent tools could be used in teaching and learning. Statistics plays vital roles in educational research, helping to understand student performance, identify trends, and evaluate the effectiveness of educational interventions. While statistical literacy, enabling individuals to critically evaluate information and make informed decisions. This paper outlines how data science and AI could be integrated into statistics education; Its impact to improve teaching and learning outcomes; addresses challenges, ethical and policy implications of integrating these technologies into statistics education.
    Keywords: Statistics, Statistics Education, Data Science, Artificial Intelligence
    Date: 2025
    URL: https://d.repec.org/n?u=RePEc:zbw:esrepo:333610
  29. By: Marcello Esposito
    Abstract: Since its inception, the Efficient Market Hypothesis (EMH) has faced persistent challenges, as numerous anomalies - such as volatility clustering, excessive trading volumes, and herding behaviour - exposed gaps between theoretical predictions and actual market dynamics. In response, economists developed alternative frameworks that relaxed EMH’s strict assumptions, distinguishing between different types of investors (e.g., “chartists†and “fundamentalists†) and incorporating bounded rationality, learning, and adaptation. This line of research gave rise to agent-based models, which conceptualize financial markets as adaptive ecosystems and rely on simulations to capture investor interactions and the evolution of trading strategies. This paper reviews central modelling choices - such as the definition of investor heterogeneity, the specification of preferences, the mechanisms of price formation, and the processes of strategy selection - and discusses their implications for balancing realism with the complexity of calibration.
    Date: 2025–11
    URL: https://d.repec.org/n?u=RePEc:liu:liucec:2025-19
  30. By: Marek Adamczyk; Micha{\l} D\k{a}browski
    Abstract: We study a systematic approach to a popular Statistical Arbitrage technique: Pairs Trading. Instead of relying on two highly correlated assets, we replace the second asset with a replication of the first using risk factor representations. These factors are obtained through Principal Components Analysis (PCA), exchange traded funds (ETFs), and, as our main contribution, Long Short Term Memory networks (LSTMs). Residuals between the main asset and its replication are examined for mean reversion properties, and trading signals are generated for sufficiently fast mean reverting portfolios. Beyond introducing a deep learning based replication method, we adapt the framework of Avellaneda and Lee (2008) to the Polish market. Accordingly, components of WIG20, mWIG40, and selected sector indices replace the original S&P500 universe, and market parameters such as the risk free rate and transaction costs are updated to reflect local conditions. We outline the full strategy pipeline: risk factor construction, residual modeling via the Ornstein Uhlenbeck process, and signal generation. Each replication technique is described together with its practical implementation. Strategy performance is evaluated over two periods: 2017-2019 and the recessive year 2020. All methods yield profits in 2017-2019, with PCA achieving roughly 20 percent cumulative return and an annualized Sharpe ratio of up to 2.63. Despite multiple adaptations, our conclusions remain consistent with those of the original paper. During the COVID-19 recession, only the ETF based approach remains profitable (about 5 percent annual return), while PCA and LSTM methods underperform. LSTM results, although negative, are promising and indicate potential for future optimization.
    Date: 2025–11
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2512.02037
  31. By: Zhongjie Jiang
    Abstract: Although synthetic data is widely promoted as a remedy, its prevailing production paradigm -- one optimizing for statistical smoothness -- systematically removes the long-tail, cognitively grounded irregularities that characterize human text. Prolonged training on such statistically optimal but cognitively impoverished data accelerates model collapse. This paper proposes a paradigm shift: instead of imitating the surface properties of data, we simulate the cognitive processes that generate human text. We introduce the Prompt-driven Cognitive Computing Framework (PMCSF), whose core consists of a Cognitive State Decoder (CSD) that reverse-engineers unstructured text into structured cognitive vectors, and a Cognitive Text Encoder (CTE) that re-materializes these states into text enriched with human-typical imperfections via mathematically defined Cognitive Perturbation Operators. The framework is validated through a two-stage objective evaluation pipeline. First, in cognitive codec verification, CTE text yields a Jensen-Shannon divergence of 0.0614 from human text (vs. 0.4431 for standard LLM output), passes double-blind professional media review, and achieves an intraclass correlation coefficient ICC > 0.9 for cognitive profile alignment across heterogeneous models. Second, in functional gain evaluation, isomorphic stress tests in the A-share market show that strategies incorporating CTE-generated data reduce maximum drawdown by 47.4% during the 2015 crash and deliver 8.6% Defensive Alpha, exceeding transaction costs by a factor of 33. Our findings demonstrate that modelling human cognitive limitations -- not copying surface data -- enables synthetic data with genuine functional gain, offering a viable technical pathway toward resolving the AI data-collapse crisis.
    Date: 2025–12
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2512.01354
  32. By: Sam Ganzfried
    Abstract: We present an algorithm for computing all evolutionarily stable strategies in nondegenerate normal-form games with three or more players.
    Date: 2025–11
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2511.20859
  33. By: Gilles Zumbach
    Abstract: For long term investments, model portfolios are defined at the level of indexes, a setup known as Strategic Asset Allocation (SAA). The possible outcomes at a scale of a few decades can be obtained by Monte Carlo simulations, resulting in a probability density for the possible portfolio values at the investment horizon. Such studies are critical for long term wealth plannings, for example in the financial component of social insurances or in accumulated capital for retirement. The quality of the results depends on two inputs: the process used for the simulations and its parameters. The base model is a constant drift, a constant covariance and normal innovations, as pioneered by Bachelier. Beyond this model, this document presents in details a multivariate process that incorporate the most recent advances in the models for financial time series. This includes the negative correlations of the returns at a scale of a few years, the heteroskedasticity (i.e. the volatility' dynamics), and the fat tails and asymmetry for the distributions of returns. For the parameters, the quantitative outcomes depend critically on the estimate for the drift, because this is a non random contribution acting at each time step. Replacing the point forecast by a probabilistic forecast allows us to analyze the impact of the drift values, and then to incorporate this uncertainty in the Monte Carlo simulations.
    Date: 2025–11
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2511.18125
  34. By: Charlie Joyez (Université Côte d'Azur, CNRS, GREDEG, France)
    Abstract: We introduce complexity, a Stata command available on SSC that computes generalized complexity indices for specialization matrices. Originally developed for assessing economic complexity with global trade data (Hidalgo and Hausmann, 2009), these metrics have since been extended to various domains including regional development, innovation, and labor economics. The complexity command implements three core methodologies: the eigenvector method (Hausmann et al., 2011a), the Method of Reflection (Hidalgo and Hausmann, 2009), and the fitness-complexity approach (Tacchella et al., 2012). It also computes relatedness metrics such as coherence, adjacency matrices of the activity space network, and the complexity outlook as a measure of complexity potential. We describe the syntax and options, review the underlying algorithms, and provide applied examples
    Date: 2025–12
    URL: https://d.repec.org/n?u=RePEc:gre:wpaper:2025-50

This nep-cmp issue is ©2025 by Stan Miles. It is provided as is without any express or implied warranty. It may be freely redistributed in whole or in part for any purpose. If distributed in part, please include this notice.
General information on the NEP project can be found at https://nep.repec.org. For comments please write to the director of NEP, Marco Novarese at <director@nep.repec.org>. Put “NEP” in the subject, otherwise your mail may be rejected.
NEP’s infrastructure is sponsored by the School of Economics and Finance of Massey University in New Zealand.