nep-ain New Economics Papers
on Artificial Intelligence
Issue of 2026–02–02
24 papers chosen by
Ben Greiner, Wirtschaftsuniversität Wien


  1. Vibe Coding Kills Open Source By Mikl\'os Koren; G\'abor B\'ek\'es; Julian Hinz; Aaron Lohmann
  2. AI Skills Improve Job Prospects: Causal Evidence from a Hiring Experiment By Fabian Stephany; Ole Teutloff; Angelo Leone
  3. Artificial Intelligence and Skills: Evidence from Contrastive Learning in Online Job Vacancies By Hangyu Chen; Yongming Sun; Yiming Yuan
  4. AI adoption, productivity and employment: evidence from European firms By Iñaki Aldasoro; Leonardo Gambacorta; Rozalia Pal; Debora Revoltella; Christoph Weiss; Marcin Wolski
  5. The Economics of Digital Intelligence Capital: Endogenous Depreciation and the Structural Jevons Paradox By Yukun Zhang; Tianyang Zhang
  6. Generative AI as a Non-Convex Supply Shock: Market Bifurcation and Welfare Analysis By Yukun Zhang; Tianyang Zhang
  7. Strategic Delegation of Moral Decisions to AI By Tontrup, Stephan; Sprigman, Christopher Jon
  8. Normative Equivalence in human-AI Cooperation: Behaviour, Not Identity, Drives Cooperation in Mixed-Agent Groups By Nico Mutzner; Taha Yasseri; Heiko Rauhut
  9. Do people expect different behavior from large language models acting on their behalf? Evidence from norm elicitations in two canonical economic games By Pawe{\l} Niszczota; Elia Antoniou
  10. How Human is AI? Examining the Impact of Emotional Prompts on Artificial and Human and Responsiveness By Florence Bernays; Marco Henriques Pereira; Jochen Menges
  11. Large Language Models Polarize Ideologically but Moderate Affectively in Online Political Discourse By Gavin Wang; Srinaath Anbudurai; Oliver Sun; Xitong Li; Lynn Wu
  12. Large language models can effectively convince people to believe conspiracies By Thomas H. Costello; Kellin Pelrine; Matthew Kowal; Antonio A. Arechar; Jean-Fran\c{c}ois Godbout; Adam Gleave; David Rand; Gordon Pennycook
  13. Fake Date Tests: Can We Trust In-sample Accuracy of LLMs in Macroeconomic Forecasting? By Alexander Eliseev; Sergei Seleznev
  14. PriceSeer: Evaluating Large Language Models in Real-Time Stock Prediction By Bohan Liang; Zijian Chen; Qi Jia; Kaiwei Zhang; Kaiyuan Ji; Guangtao Zhai
  15. Can Large Language Models Improve Venture Capital Exit Timing After IPO? By Mohammadhossien Rashidi
  16. Look-Ahead-Bench: a Standardized Benchmark of Look-ahead Bias in Point-in-Time LLMs for Finance By Mostapha Benhenda
  17. AI Cap-and-Trade: Efficiency Incentives for Accessibility and Sustainability By Marco Bornstein; Amrit Singh Bedi
  18. Women Worry, Men Adopt: How Gendered Perceptions Shape the Use of Generative AI By Fabian Stephany; Jedrzej Duszynski
  19. When Is Self-Disclosure Optimal? Incentives and Governance of AI-Generated Content By Juan Wu; Zhe; Zhang; Amit Mehra
  20. Artificial Intelligence and the US Economy: An Accounting Perspective on Investment and Production By Luisa Carpinelli; Filippo Natoli; Marco Taboga
  21. Human-AI Collaboration in Radiology: The Case of Pulmonary Embolism By Paul Goldsmith-Pinkham; Chenhao Tan; Alexander K. Zentefis
  22. Autonomous Market Intelligence: Agentic AI Nowcasting Predicts Stock Returns By Zefeng Chen; Darcy Pu
  23. All That Glisters Is Not Gold: A Benchmark for Reference-Free Counterfactual Financial Misinformation Detection By Yuechen Jiang; Zhiwei Liu; Yupeng Cao; Yueru He; Chen Xu; Ziyang Xu; Zhiyang Deng; Prayag Tiwari; Xi Chen; Alejandro Lopez-Lira; Jimin Huang; Junichi Tsujii; Sophia Ananiadou
  24. A Model of Artificial Jagged Intelligence By Joshua S. Gans

  1. By: Mikl\'os Koren; G\'abor B\'ek\'es; Julian Hinz; Aaron Lohmann
    Abstract: Generative AI is changing how software is produced and used. In vibe coding, an AI agent builds software by selecting and assembling open-source software (OSS), often without users directly reading documentation, reporting bugs, or otherwise engaging with maintainers. We study the equilibrium effects of vibe coding on the OSS ecosystem. We develop a model with endogenous entry and heterogeneous project quality in which OSS is a scalable input into producing more software. Users choose whether to use OSS directly or through vibe coding. Vibe coding raises productivity by lowering the cost of using and building on existing code, but it also weakens the user engagement through which many maintainers earn returns. When OSS is monetized only through direct user engagement, greater adoption of vibe coding lowers entry and sharing, reduces the availability and quality of OSS, and reduces welfare despite higher productivity. Sustaining OSS at its current scale under widespread vibe coding requires major changes in how maintainers are paid.
    Date: 2026–01
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2601.15494
  2. By: Fabian Stephany; Ole Teutloff; Angelo Leone
    Abstract: The growing adoption of artificial intelligence (AI) technologies has heightened interest in the labour market value of AI-related skills, yet causal evidence on their role in hiring decisions remains scarce. This study examines whether AI skills serve as a positive hiring signal and whether they can offset conventional disadvantages such as older age or lower formal education. We conduct an experimental survey with 1, 700 recruiters from the United Kingdom and the United States. Using a paired conjoint design, recruiters evaluated hypothetical candidates represented by synthetically designed resumes. Across three occupations - graphic designer, office assistant, and software engineer - AI skills significantly increase interview invitation probabilities by approximately 8 to 15 percentage points. AI skills also partially or fully offset disadvantages related to age and lower education, with effects strongest for office assistants, where formal AI certification plays an additional compensatory role. Effects are weaker for graphic designers, consistent with more skeptical recruiter attitudes toward AI in creative work. Finally, recruiters' own background and AI usage significantly moderate these effects. Overall, the findings demonstrate that AI skills function as a powerful hiring signal and can mitigate traditional labour market disadvantages, with implications for workers' skill acquisition strategies and firms' recruitment practices.
    Date: 2026–01
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2601.13286
  3. By: Hangyu Chen; Yongming Sun; Yiming Yuan
    Abstract: We investigate the impact of artificial intelligence (AI) adoption on skill requirements using 14 million online job vacancies from Chinese listed firms (2018-2022). Employing a novel Extreme Multi-Label Classification (XMLC) algorithm trained via contrastive learning and LLM-driven data augmentation, we map vacancy requirements to the ESCO framework. By benchmarking occupation-skill relationships against 2018 O*NET-ESCO mappings, we document a robust causal relationship between AI adoption and the expansion of skill portfolios. Our analysis identifies two distinct mechanisms. First, AI reduces information asymmetry in the labor market, enabling firms to specify current occupation-specific requirements with greater precision. Second, AI empowers firms to anticipate evolving labor market dynamics. We find that AI adoption significantly increases the demand for "forward-looking" skills--those absent from 2018 standards but subsequently codified in 2022 updates. This suggests that AI allows firms to lead, rather than follow, the formal evolution of occupational standards. Our findings highlight AI's dual role as both a stabilizer of current recruitment information and a catalyst for proactive adaptation to future skill shifts.
    Date: 2026–01
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2601.03558
  4. By: Iñaki Aldasoro; Leonardo Gambacorta; Rozalia Pal; Debora Revoltella; Christoph Weiss; Marcin Wolski
    Abstract: This paper provides new evidence on how the adoption of artificial intelligence (AI) affects productivity and employment in Europe. Using matched EIBIS-ORBIS data on more than 12, 000 non-financial firms in the European Union (EU) and United States (US), we instrument the adoption of AI by EU firms by assigning the adoption rates of US peers to isolate exogenous technological exposure. Our results show that AI adoption increases the level of labor productivity by 4%. Productivity gains are due to capital deepening, as we find no adverse effects on firm-level employment. This suggests that AI increases worker output rather than replacing labor in the short run, though longer-term effects remain uncertain. However, productivity benefits of AI adoption are unevenly distributed and concentrate in medium and large firms. Moreover, AI-adopting firms are more innovative and their workers earn higher wages. Our analysis also highlights the critical role of complementary investments in software and data or workforce training to fully unlock the productivity gains of AI adoption.
    Keywords: artificial intelligence, firm productivity, Europe, digital transformation
    JEL: D22 J24 L25 O33 O47
    Date: 2026–01
    URL: https://d.repec.org/n?u=RePEc:bis:biswps:1325
  5. By: Yukun Zhang; Tianyang Zhang
    Abstract: This paper develops a micro-founded economic theory of the AI industry by modeling large language models as a distinct asset class-Digital Intelligence Capital-characterized by data-compute complementarities, increasing returns to scale, and relative (rather than absolute) valuation. We show that these features fundamentally reshape industry dynamics along three dimensions. First, because downstream demand depends on relative capability, innovation by one firm endogenously depreciates the economic value of rivals' existing capital, generating a persistent innovation pressure we term the Red Queen Effect. Second, falling inference prices induce downstream firms to adopt more compute-intensive agent architectures, rendering aggregate demand for compute super-elastic and producing a structural Jevons paradox. Third, learning from user feedback creates a data flywheel that can destabilize symmetric competition: when data accumulation outpaces data decay, the market bifurcates endogenously toward a winner-takes-all equilibrium. We further characterize conditions under which expanding upstream capabilities erode downstream application value (the Wrapper Trap). A calibrated agent-based model confirms these mechanisms and their quantitative implications. Together, the results provide a unified framework linking intelligence production upstream with agentic demand downstream, offering new insights into competition, scalability, and regulation in the AI economy.
    Date: 2026–01
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2601.12339
  6. By: Yukun Zhang; Tianyang Zhang
    Abstract: The diffusion of Generative AI (GenAI) constitutes a supply shock of a fundamentally different nature: while marginal production costs approach zero, content generation creates congestion externalities through information pollution. We develop a three-layer general equilibrium framework to study how this non-convex technology reshapes market structure, transition dynamics, and social welfare. In a static vertical differentiation model, we show that the GenAI cost shock induces a kinked production frontier that bifurcates the market into exit, AI, and human segments, generating a ``middle-class hollow'' in the quality distribution. To analyze adjustment paths, we embed this structure in a mean-field evolutionary system and a calibrated agent-based model with bounded rationality. The transition to the AI-integrated equilibrium is non-monotonic: rather than smooth diffusion, the economy experiences a temporary ecological collapse driven by search frictions and delayed skill adaptation, followed by selective recovery. Survival depends on asymmetric skill reconfiguration, whereby humans retreat from technical execution toward semantic creativity. Finally, we show that the welfare impact of AI adoption is highly sensitive to pollution intensity: low congestion yields monotonic welfare gains, whereas high pollution produces an inverted-U relationship in which further AI expansion reduces total welfare. These results imply that laissez-faire adoption can be inefficient and that optimal governance must shift from input regulation toward output-side congestion management.
    Date: 2026–01
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2601.12488
  7. By: Tontrup, Stephan; Sprigman, Christopher Jon
    Abstract: Our study examines how individuals perceive the moral agency of artificial intelligence (AI), and, specifically, whether individuals believe that by involving AI as their agent, they can offload to the AI some of their responsibility for a morally sensitive decision. Existing literature shows that people often delegate self-interested decisions to human agents to mitigate their moral responsibility for unethical outcomes. This research explores whether individuals will similarly delegate such decisions to AI to reduce moral costs. Our study shows that many individuals perceive the AI as capable of assuming moral responsibility. These individuals delegate to the AI and delegating leads them to act more assertively in their self-interest while experiencing lower moral costs. Participants (hereinafter, "Allocators") took part in a dictator game, allocating a $10 endowment between themselves and a Recipient. In the experimental treatment, Allocators could involve ChatGPT in their allocation decision, at the cost of incurring added time to complete the experiment. When engaged, the AI executed the transfer by informing the Recipient of a necessary payment code. Around 35% of Allocators chose to involve the AI, despite the opportunity costs of a much-prolonged process. To isolate the effect of the AI's perceived responsibility, a control condition replaced the AI with a non-agentive computer program, while maintaining identical decision protocols. This design controlled for factors such as social distance and substantive influence by the AI. Allocators who involved the AI transferred significantly less money to the Recipient, suggesting that delegating the transfer to AI reduced the moral costs associated with self-interested decisions. This is supported by the fact that prosocial individuals, who face higher moral costs from violating a norm and thus would without delegation transfer more than proself individuals, were significantly more likely to involve the AI. A responsibility measure indicates that Allocators who attributed more responsibility for the transfer to the AI were also more likely to involve the AI. The study suggests that AI systems provide human actors with an easily accessible, low-cost, and hard-to-monitor means of offloading personal moral responsibility, highlighting the need to consider in AI regulation not only the inherent risks of AI output, but also how AI's perceived moral agency can influence human behavior and ethical accountability in human-AI interaction.
    Keywords: AI, Delegation, Moral Outsourcing, Prosociality
    JEL: C91 D91 O33 D63
    Date: 2025
    URL: https://d.repec.org/n?u=RePEc:zbw:esprep:335206
  8. By: Nico Mutzner; Taha Yasseri; Heiko Rauhut
    Abstract: The introduction of artificial intelligence (AI) agents into human group settings raises essential questions about how these novel participants influence cooperative social norms. While previous studies on human-AI cooperation have primarily focused on dyadic interactions, little is known about how integrating AI agents affects the emergence and maintenance of cooperative norms in small groups. This study addresses this gap through an online experiment using a repeated four-player Public Goods Game (PGG). Each group consisted of three human participants and one bot, which was framed either as human or AI and followed one of three predefined decision strategies: unconditional cooperation, conditional cooperation, or free-riding. In our sample of 236 participants, we found that reciprocal group dynamics and behavioural inertia primarily drove cooperation. These normative mechanisms operated identically across conditions, resulting in cooperation levels that did not differ significantly between human and AI labels. Furthermore, we found no evidence of differences in norm persistence in a follow-up Prisoner's Dilemma, or in participants' normative perceptions. Participants' behaviour followed the same normative logic across human and AI conditions, indicating that cooperation depended on group behaviour rather than partner identity. This supports a pattern of normative equivalence, in which the mechanisms that sustain cooperation function similarly in mixed human-AI and all human groups. These findings suggest that cooperative norms are flexible enough to extend to artificial agents, blurring the boundary between humans and AI in collective decision-making.
    Date: 2026–01
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2601.20487
  9. By: Pawe{\l} Niszczota; Elia Antoniou
    Abstract: While delegating tasks to large language models (LLMs) can save people time, there is growing evidence that offloading tasks to such models produces social costs. We use behavior in two canonical economic games to study whether people have different expectations when decisions are made by LLMs acting on their behalf instead of themselves. More specifically, we study the social appropriateness of a spectrum of possible behaviors: when LLMs divide resources on our behalf (Dictator Game and Ultimatum Game) and when they monitor the fairness of splits of resources (Ultimatum Game). We use the Krupka-Weber norm elicitation task to detect shifts in social appropriateness ratings. Results of two pre-registered and incentivized experimental studies using representative samples from the UK and US (N = 2, 658) show three key findings. First, people find that offers from machines - when no acceptance is necessary - are judged to be less appropriate than when they come from humans, although there is no shift in the modal response. Second - when acceptance is necessary - it is more appropriate for a person to reject offers from machines than from humans. Third, receiving a rejection of an offer from a machine is no less socially appropriate than receiving the same rejection from a human. Overall, these results suggest that people apply different norms for machines deciding on how to split resources but are not opposed to machines enforcing the norms. The findings are consistent with offers made by machines now being viewed as having both a cognitive and emotional component.
    Date: 2026–01
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2601.15312
  10. By: Florence Bernays (University of Zurich); Marco Henriques Pereira (University of Zurich); Jochen Menges (University of Zurich)
    Abstract: This research examines how the emotional tone of human-AI interactions shapes ChatGPT and human behavior. In a between-subject experiment, we asked participants to express a specific emotion while working with ChatGPT (GPT-4.0) on two tasks, including writing a public response and addressing an ethical dilemma. We found that compared to interactions where participants maintained a neutral tone, ChatGPT showed greater improvement in its answers when participants praised ChatGPT for its responses. Expressing anger towards ChatGPT also led to a higher albeit smaller improvement relative to the neutral condition, whereas blaming ChatGPT did not improve its answers. When addressing an ethical dilemma, ChatGPT prioritized corporate interests less when participants expressed anger towards it, while blaming increases its emphasis on protecting the public interest. Additionally, we found that people used more negative, hostile, and disappointing expressions in human-human communication after interactions during which participants blamed rather than praised for their responses. Together, our findings demonstrate that the emotional tone people apply in human-AI interactions not only shape ChatGPT's outputs but also carry over into subsequent human-human communication.
    Date: 2026–01
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2601.05104
  11. By: Gavin Wang; Srinaath Anbudurai; Oliver Sun; Xitong Li; Lynn Wu
    Abstract: The emergence of large language models (LLMs) is reshaping how people engage in political discourse online. We examine how the release of ChatGPT altered ideological and emotional patterns in the largest political forum on Reddit. Analysis of millions of comments shows that ChatGPT intensified ideological polarization: liberals became more liberal, and conservatives more conservative. This shift does not stem from the creation of more persuasive or ideologically extreme original content using ChatGPT. Instead, it originates from the tendency of ChatGPT-generated comments to echo and reinforce the viewpoint of original posts, a pattern consistent with algorithmic sycophancy. Yet, despite growing ideological divides, affective polarization, measured by hostility and toxicity, declined. These findings reveal that LLMs can simultaneously deepen ideological separation and foster more civil exchanges, challenging the long-standing assumption that extremity and incivility necessarily move together.
    Date: 2026–01
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2601.20238
  12. By: Thomas H. Costello; Kellin Pelrine; Matthew Kowal; Antonio A. Arechar; Jean-Fran\c{c}ois Godbout; Adam Gleave; David Rand; Gordon Pennycook
    Abstract: Large language models (LLMs) have been shown to be persuasive across a variety of context. But it remains unclear whether this persuasive power advantages truth over falsehood, or if LLMs can promote misbeliefs just as easily as refuting them. Here, we investigate this question across three pre-registered experiments in which participants (N = 2, 724 Americans) discussed a conspiracy theory they were uncertain about with GPT-4o, and the model was instructed to either argue against ("debunking") or for ("bunking") that conspiracy. When using a "jailbroken" GPT-4o variant with guardrails removed, the AI was as effective at increasing conspiracy belief as decreasing it. Concerningly, the bunking AI was rated more positively, and increased trust in AI, more than the debunking AI. Surprisingly, we found that using standard GPT-4o produced very similar effects, such that the guardrails imposed by OpenAI did little to revent the LLM from promoting conspiracy beliefs. Encouragingly, however, a corrective conversation reversed these newly induced conspiracy beliefs, and simply prompting GPT-4o to only use accurate information dramatically reduced its ability to increase conspiracy beliefs. Our findings demonstrate that LLMs possess potent abilities to promote both truth and falsehood, but that potential solutions may exist to help mitigate this risk.
    Date: 2026–01
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2601.05050
  13. By: Alexander Eliseev; Sergei Seleznev
    Abstract: Large language models (LLMs) are a type of machine learning tool that economists have started to apply in their empirical research. One such application is macroeconomic forecasting with backtesting of LLMs, even though they are trained on the same data that is used to estimate their forecasting performance. Can these in-sample accuracy results be extrapolated to the model's out-of-sample performance? To answer this question, we developed a family of prompt sensitivity tests and two members of this family, which we call the fake date tests. These tests aim to detect two types of biases in LLMs' in-sample forecasts: lookahead bias and context bias. According to the empirical results, none of the modern LLMs tested in this study passed our first test, signaling the presence of lookahead bias in their in-sample forecasts.
    Date: 2026–01
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2601.07992
  14. By: Bohan Liang; Zijian Chen; Qi Jia; Kaiwei Zhang; Kaiyuan Ji; Guangtao Zhai
    Abstract: Stock prediction, a subject closely related to people's investment activities in fully dynamic and live environments, has been widely studied. Current large language models (LLMs) have shown remarkable potential in various domains, exhibiting expert-level performance through advanced reasoning and contextual understanding. In this paper, we introduce PriceSeer, a live, dynamic, and data-uncontaminated benchmark specifically designed for LLMs performing stock prediction tasks. Specifically, PriceSeer includes 110 U.S. stocks from 11 industrial sectors, with each containing 249 historical data points. Our benchmark implements both internal and external information expansion, where LLMs receive extra financial indicators, news, and fake news to perform stock price prediction. We evaluate six cutting-edge LLMs under different prediction horizons, demonstrating their potential in generating investment strategies after obtaining accurate price predictions for different sectors. Additionally, we provide analyses of LLMs' suboptimal performance in long-term predictions, including the vulnerability to fake news and specific industries. The code and evaluation data will be open-sourced at https://github.com/BobLiang2113/PriceSee r.
    Date: 2025–12
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2601.06088
  15. By: Mohammadhossien Rashidi
    Abstract: Exit timing after an IPO is one of the most consequential decisions for venture capital (VC) investors, yet existing research focuses mainly on describing when VCs exit rather than evaluating whether those choices are economically optimal. Meanwhile, large language models (LLMs) have shown promise in synthesizing complex financial data and textual information but have not been applied to post-IPO exit decisions. This study introduces a framework that uses LLMs to estimate the optimal time for VC exit by analyzing monthly post IPO information financial performance, filings, news, and market signals and recommending whether to sell or continue holding. We compare these LLM generated recommendations with the actual exit dates observed for VCs and compute the return differences between the two strategies. By quantifying gains or losses associated with following the LLM, this study provides evidence on whether AI-driven guidance can improve exit timing and complements traditional hazard and real-options models in venture capital research.
    Date: 2025–12
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2601.00810
  16. By: Mostapha Benhenda (LAGA)
    Abstract: We introduce Look-Ahead-Bench, a standardized benchmark measuring look-ahead bias in Point-in-Time (PiT) Large Language Models (LLMs) within realistic and practical financial workflows. Unlike most existing approaches that primarily test inner lookahead knowledge via Q\\&A, our benchmark evaluates model behavior in practical scenarios. To distinguish genuine predictive capability from memorization-based performance, we analyze performance decay across temporally distinct market regimes, incorporating several quantitative baselines to establish performance thresholds. We evaluate prominent open-source LLMs -- Llama 3.1 (8B and 70B) and DeepSeek 3.2 -- against a family of Point-in-Time LLMs (Pitinf-Small, Pitinf-Medium, and frontier-level model Pitinf-Large) from PiT-Inference. Results reveal significant lookahead bias in standard LLMs, as measured with alpha decay, unlike Pitinf models, which demonstrate improved generalization and reasoning abilities as they scale in size. This work establishes a foundation for the standardized evaluation of temporal bias in financial LLMs and provides a practical framework for identifying models suitable for real-world deployment. Code is available on GitHub: https://github.com/benstaf/lookaheadbenc h
    Date: 2026–01
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2601.13770
  17. By: Marco Bornstein; Amrit Singh Bedi
    Abstract: The race for artificial intelligence (AI) dominance often prioritizes scale over efficiency. Hyper-scaling is the common industry approach: larger models, more data, and as many computational resources as possible. Using more resources is a simpler path to improved AI performance. Thus, efficiency has been de-emphasized. Consequently, the need for costly computational resources has marginalized academics and smaller companies. Simultaneously, increased energy expenditure, due to growing AI use, has led to mounting environmental costs. In response to accessibility and sustainability concerns, we argue for research into, and implementation of, market-based methods that incentivize AI efficiency. We believe that incentivizing efficient operations and approaches will reduce emissions while opening new opportunities for academics and smaller companies. As a call to action, we propose a cap-and-trade system for AI. Our system provably reduces computations for AI deployment, thereby lowering emissions and monetizing efficiency to the benefit of of academics and smaller companies.
    Date: 2026–01
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2601.19886
  18. By: Fabian Stephany; Jedrzej Duszynski
    Abstract: Generative artificial intelligence (GenAI) is diffusing rapidly, yet its adoption is strikingly unequal. Using nationally representative UK survey data from 2023 to 2024, we show that women adopt GenAI substantially less often than men because they perceive its societal risks differently. We construct a composite index capturing concerns about mental health, privacy, climate impact, and labor market disruption. This index explains between 9 and 18 percent of the variation in GenAI adoption and ranks among the strongest predictors for women across all age groups, surpassing digital literacy and education for young women. Intersectional analyses show that the largest disparities arise among younger, digitally fluent individuals with high societal risk concerns, where gender gaps in personal use exceed 45 percentage points. Using a synthetic twin panel design, we show that increased optimism about AI's societal impact raises GenAI use among young women from 13 percent to 33 percent, substantially narrowing the gender divide. These findings indicate that gendered perceptions of AI's social and ethical consequences, rather than access or capability, are the primary drivers of unequal GenAI adoption, with implications for productivity, skill formation, and economic inequality in an AI enabled economy.
    Date: 2026–01
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2601.03880
  19. By: Juan Wu (James); Zhe (James); Zhang; Amit Mehra
    Abstract: Generative artificial intelligence (Gen-AI) is reshaping content creation on digital platforms by reducing production costs and enabling scalable output of varying quality. In response, platforms have begun adopting disclosure policies that require creators to label AI-generated content, often supported by imperfect detection and penalties for non-compliance. This paper develops a formal model to study the economic implications of such disclosure regimes. We compare a non-disclosure benchmark, in which the platform alone detects AI usage, with a mandatory self-disclosure regime in which creators strategically choose whether to disclose or conceal AI use under imperfect enforcement. The model incorporates heterogeneous creators, viewer discounting of AI-labeled content, trust penalties following detected non-disclosure, and endogenous enforcement. The analysis shows that disclosure is optimal only when both the value of AI-generated content and its cost-saving advantage are intermediate. As AI capability improves, the platform's optimal enforcement strategy evolves from strict deterrence to partial screening and eventual deregulation. While disclosure reliably increases transparency, it reduces aggregate creator surplus and can suppress high-quality AI content when AI is technologically advanced. Overall, the results characterize disclosure as a strategic governance instrument whose effectiveness depends on technological maturity and trust frictions.
    Date: 2026–01
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2601.18654
  20. By: Luisa Carpinelli; Filippo Natoli; Marco Taboga
    Abstract: Artificial intelligence (AI) has moved to the center of policy, market, and academic debates, but its macroeconomic footprint is still only partly understood. This paper provides an overview on how the current AI wave is captured in US national accounts, combining a simple macro-accounting framework with a stylized description of the AI production process. We highlight the crucial role played by data centers, which constitute the backbone of the AI ecosystem and have attracted formidable investment in 2025, as they are indispensable for meeting the rapidly increasing worldwide demand for AI services. We document that the boom in IT and AI-related capital expenditure in the first three quarters of the year has given an outsized boost to aggregate demand, while its contribution to GDP growth is smaller once the high import content of AI hardware is netted out. Furthermore, simple calculations suggest that, at current utilization rates and pricing, the production of services originating in new AI data centers could contribute to GDP over the turn of the next quarters on a scale comparable to that of investment spending to date. Short reinvestment cycles and uncertainty about future AI demand, while not currently acting as a macroeconomic drag, can nevertheless fuel macroeconomic risks over the medium term.
    Date: 2026–01
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2601.11196
  21. By: Paul Goldsmith-Pinkham; Chenhao Tan; Alexander K. Zentefis
    Abstract: We study how radiologists use AI to diagnose pulmonary embolism (PE), tracking over 100, 000 scans interpreted by nearly 400 radiologists during the staggered rollout of a real-world FDA-approved diagnostic platform in a hospital system. When AI flags PE, radiologists agree 84% of the time; when AI predicts no PE, they agree 97%. Disagreement evolves substantially: radiologists initially reject AI-positive PEs in 30% of cases, dropping to 12% by year two. Despite a 16% increase in scan volume, diagnostic speed remains stable while per-radiologist monthly volumes nearly double, with no change in patient mortality -- suggesting AI improves workflow without compromising outcomes. We document significant heterogeneity in AI collaboration: some radiologists reject AI-flagged PEs half the time while others accept nearly always; female radiologists are 6 percentage points less likely to override AI than male radiologists. Moderate AI engagement is associated with the highest agreement, whereas both low and high engagement show more disagreement. Follow-up imaging reveals that when radiologists override AI to diagnose PE, 54% of subsequent scans show both agreeing on no PE within 30 days.
    Date: 2026–01
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2601.13379
  22. By: Zefeng Chen; Darcy Pu
    Abstract: Can fully agentic AI nowcast stock returns? We deploy a state-of-the-art Large Language Model to evaluate the attractiveness of each Russell 1000 stock daily, starting from April 2025 when AI web interfaces enabled real-time search. Our data contribution is unique along three dimensions. First, the nowcasting framework is completely out-of-sample and free of look-ahead bias by construction: predictions are collected at the current edge of time, ensuring the AI has no knowledge of future outcomes. Second, this temporal design is irreproducible -- once the information environment passes, it can never be recreated. Third, our framework is 100% agentic: we do not feed the model news, disclosures, or curated text; it autonomously searches the web, filters sources, and synthesises information into quantitative predictions. We find that AI possesses genuine stock selection ability, but only for identifying top winners. Longing the 20 highest-ranked stocks generates a daily Fama-French five-factor plus momentum alpha of 18.4 basis points and an annualised Sharpe ratio of 2.43. Critically, these returns derive from an implementable strategy trading highly liquid Russell 1000 constituents, with transaction costs representing less than 10\% of gross alpha. However, this predictability is highly concentrated: expanding beyond the top tier rapidly dilutes alpha, and bottom-ranked stocks exhibit returns statistically indistinguishable from the market. We hypothesise that this asymmetry reflects online information structure: genuinely positive news generates coherent signals, while negative news is contaminated by strategic corporate obfuscation and social media noise.
    Date: 2026–01
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2601.11958
  23. By: Yuechen Jiang; Zhiwei Liu; Yupeng Cao; Yueru He; Chen Xu; Ziyang Xu; Zhiyang Deng; Prayag Tiwari; Xi Chen; Alejandro Lopez-Lira; Jimin Huang; Junichi Tsujii; Sophia Ananiadou
    Abstract: We introduce RFC Bench, a benchmark for evaluating large language models on financial misinformation under realistic news. RFC Bench operates at the paragraph level and captures the contextual complexity of financial news where meaning emerges from dispersed cues. The benchmark defines two complementary tasks: reference free misinformation detection and comparison based diagnosis using paired original perturbed inputs. Experiments reveal a consistent pattern: performance is substantially stronger when comparative context is available, while reference free settings expose significant weaknesses, including unstable predictions and elevated invalid outputs. These results indicate that current models struggle to maintain coherent belief states without external grounding. By highlighting this gap, RFC Bench provides a structured testbed for studying reference free reasoning and advancing more reliable financial misinformation detection in real world settings.
    Date: 2026–01
    URL: https://d.repec.org/n?u=RePEc:arx:papers:2601.04160
  24. By: Joshua S. Gans
    Abstract: Generative AI systems often display highly uneven performance across tasks that appear “nearby”: they can be excellent on one prompt and confidently wrong on another with only small changes in wording or context. We call this phenomenon Artificial Jagged Intelligence (AJI). This paper develops a tractable economic model of AJI that treats adoption as an information problem: users care about local reliability, but typically observe only coarse, global quality signals. In a baseline one-dimensional landscape, truth is a rough Brownian process, and the model “knows” scattered points drawn from a Poisson process. The model interpolates optimally, and the local error is measured by posterior variance. We derive an adoption threshold for a blind user, show that experienced errors are amplified by the inspection paradox, and interpret scaling laws as denser coverage that improves average quality without eliminating jaggedness. We then study mastery and calibration: a calibrated user who can condition on local uncertainty enjoys positive expected value even in domains that fail the blind adoption test. Modelling mastery as learning a reliability map via Gaussian process regression yields a learning-rate bound driven by information gain, clarifying when discovering “where the model works” is slow. Finally, we study how scaling interacts with discoverability: when calibrated signals and user mastery accelerate the harvesting of scale improvements, and when opacity can make gains from scaling effectively invisible.
    JEL: D83 O33
    Date: 2026–01
    URL: https://d.repec.org/n?u=RePEc:nbr:nberwo:34712

This nep-ain issue is ©2026 by Ben Greiner. It is provided as is without any express or implied warranty. It may be freely redistributed in whole or in part for any purpose. If distributed in part, please include this notice.
General information on the NEP project can be found at https://nep.repec.org. For comments please write to the director of NEP, Marco Novarese at <director@nep.repec.org>. Put “NEP” in the subject, otherwise your mail may be rejected.
NEP’s infrastructure is sponsored by the School of Economics and Finance of Massey University in New Zealand.