Regulating Artificial Intelligence in the EU, United States and China - Implications for energy systems

The growing prevalence and potential impact of artificial intelligence (AI) on society rises the need for regulation. In return, the shape of regulations will affect the application potential of AI across all economic sectors. This study compares the approaches to regulate AI in the European Union (EU), the United States (US) and China (CN). We then apply the findings of our comparative analysis on the energy sector, assessing the effects of each regulatory approach on the operation of a AI-based short-term electricity demand forecasting application. Our findings show that operationalizing AI applications will face very different challenges across geographies, with important implications for..

Computational Economics

Pattern Mining for Anomaly Detection in Graphs: Application to Fraud in Public Procurement

In the context of public procurement, several indicators called red flags are used to estimate fraud risk. They are computed according to certain contract attributes and are therefore dependent on the proper filling of the contract and award notices. However, these attributes are very often missing in practice, which prohibits red flags computation. Traditional fraud detection approaches focus on tabular data only, considering each contract separately, and are therefore very sensitive to this issue. In this work, we adopt a graph-based method allowing leveraging relations between contracts, to compensate for the missing attributes. We propose PANG (Pattern-Based Anomaly Detection in Graphs),..

Computational Economics

Nowcasting trade in value added indicators

Trade in value added (TiVA) indicators are increasingly used to monitor countries’ integration into global supply chains. However, they are published with a significant lag - often two or three years - which reduces their relevance for monitoring recent economic developments. This paper aims to provide more timely insights into the international fragmentation of production by exploring new ways of nowcasting five TiVA indicators for the years 2021 and 2022 covering a panel of 41 economies at the economy-wide level and for 24 industry sectors. The analysis relies on a range of models, including Gradient boosted trees (GBM), and other machine-learning techniques, in a panel setting, uses a w..

Computational Economics

Informal employment from migration shocks

We propose a new approach to detect and quantify informal employment resulting from irregular migration shocks. Focusing on a largely informal sector, agriculture, and on the exogenous variation from the Arab Spring wave on southern Italian coasts, we use machine-learning techniques to document abnormal increases in reported (vs. predicted) labor productivity on vineyards hit by the shock. Misreporting is largely heterogeneous across farms depending e.g. on size and grape quality. The shock resulted in a 6% increase in informal employment, equivalent to one undeclared worker for every three farms on average and 23, 000 workers in total over 2011-2012. Misreporting causes significant increase..

Computational Economics

Meta-Analysis of Social Science Research: A Practitioner´s Guide

This paper provides concise, nontechnical, step-by-step guidelines on how to conduct a modern meta-analysis, especially in social sciences. We treat publication bias, p-hacking, and heterogeneity as phenomena meta-analysts must always confront. To this end, we provide concrete methodological recommendations. Meta-analysis methods have advanced notably over the last few years. Yet many meta-analyses still rely on outdated approaches, some ignoring publication bias and systematic heterogeneity. While limitations persist, recently developed techniques allow robust inference even in the face of formidable problems in the underlying empirical literature. The purpose of this paper is to summarize ..

Computational Economics

The FOMC versus the Staff: Do Policymakers Add Value in Their Tales?

Using close to 40 years of textual data from FOMC transcripts and the Federal Reserve staff's Greenbook/Tealbook, we extend Romer and Romer (2008) to test if the FOMC adds information relative to its staff forecasts not via its own quantitative forecasts but via its words. We use methods from natural language processing to extract from both types of document text-based forecasts that capture attentiveness to and sentiment about the macroeconomy. We test whether these text-based forecasts provide value-added in explaining the distribution of outcomes for GDP growth, the unemployment rate, and inflation. We find that FOMC tales about macroeconomic risks do add value in the tails, especially fo..

Computational Economics

Applying Machine Learning Algorithms to Predict the Size of the Informal Economy

The use of machine learning models and techniques to predict economic variables has been growing lately, motivated by their better performance when compared to that of linear models. Although linear models have the advantage of considerable interpretive power, efforts have intensified in recent years to make machine learning models more interpretable. In this paper, tests are conducted to determine whether models based on machine learning algorithms have better performance relative to that of linear models for predicting the size of the informal economy. The paper also explores whether the determinants of such size detected as the most important by machine learning models are the same as tho..

Computational Economics

How uncertainty shapes herding in the corporate use of artificial intelligence technology

Computational Economics

A critical assessment of neural networks as meta-model of a farm optimization model

Mixed Integer programming (MIP) is frequently used in agricultural economics to solve farm-level optimization problems, but it can be computationally intensive especially when the number of binary or integer variables becomes large. In order to speed up simulations, for instance for large-scale sensitivity analysis or application to larger farm populations, meta-models can be derived from the original MIP and applied as an approximator instead. To test and assess this approach, we train Artificial Neural Networks (ANNs) as a meta-model of a farm-scale MIP model. This study compares different ANNs from various perspectives to assess to what extent they are able to replace the original MIP mod..

Computational Economics

Munging the Ghosts in the Machine: Coded Bias and the Craft of Wrangling Archival Data

Data wrangling is typically treated as an obligatory, codified, and ideally automated step in the machine learning (ML) pipeline. In contrast, we suggest that archival data wrangling is a theory-driven process best understood as a practical craft. Drawing on empirical examples from contemporary computational social science, we identify nine core modes of data wrangling, which can be seen as a sequence but are iterative and nonlinear in practice. Moreover, we discuss how data wrangling can address issues of algorithmic bias. While ML has shifted the focus towards architectural engineering, we assert that to properly engage with machine learning is to properly engage with data wrangling.

Computational Economics

Multi-level Fusion in Deep Convolutional Networks for Enhanced Image Analysis

Multi-level Fusion in Deep Convolutional Networks for Enhanced Image Analysis

Computational Economics

Estimating HANK for Central Banks

We provide a toolkit for efficient online estimation of heterogeneous agent (HA) New Keynesian (NK) models based on Sequential Monte Carlo methods. We use this toolkit to compare the out-of-sample forecasting accuracy of a prominent HANK model, Bayer et al. (2022), to that of the representative agent (RA) NK model of Smets and Wouters (2007, SW). We find that HANK’s accuracy for real activity variables is notably inferior to that of SW. The results for consumption in particular are disappointing since the main difference between RANK and HANK is the replacement of the RA Euler equation with the aggregation of individual households’ consumption policy functions, which reflects inequality.

Computational Economics

Nowcasting world trade with machine learning: a three-step approach

We nowcast world trade using machine learning, distinguishing between tree-based methods (random forest, gradient boosting) and their regression-based counterparts (macroeconomic random forest, linear gradient boosting). While much less used in the literature, the latter are found to outperform not only the tree-based techniques, but also more “traditional” linear and non-linear techniques (OLS, Markov-switching, quantile regression). They do so significantly and consistently across different horizons and real-time datasets. To further improve performances when forecasting with machine learning, we propose a flexible three-step approach composed of (step 1) pre-selection, (step 2) factor..

Computational Economics

Aggregation of Information and Communications Industry and Self-organization Simulation Using an Agent-based Model (Japanese)

As a preliminary step to conducting a self-organization simulation of the concentration and dispersion of the information and communications industry, we will quantify the spatial concentration of the information and communications industry in large cities in Japan. Spatial analysis of the information and communications industry in Sapporo, Sendai, Hiroshima, and Fukuoka, which are regional core cities, in addition to the 23 wards of Tokyo, was conducted using Chocho data from the "Economic Census." As a result of detecting spatial autocorrelation in small area units in each city, hotspots indicating concentration of information and communications business establishments were detected in the..

Computational Economics

Amortized neural networks for agent-based model forecasting

In this paper, we propose a new procedure for unconditional and conditional forecasting in agent-based models. The proposed algorithm is based on the application of amortized neural networks and consists of two steps. The first step simulates artificial datasets from the model. In the second step, a neural network is trained to predict the future values of the variables using the history of observations. The main advantage of the proposed algorithm is its speed. This is due to the fact that, after the training procedure, it can be used to yield predictions for almost any data without additional simulations or the re-estimation of the neural network

Computational Economics

Financial Fraud Detection: A Comparative Study of Quantum Machine Learning Models

In this research, a comparative study of four Quantum Machine Learning (QML) models was conducted for fraud detection in finance. We proved that the Quantum Support Vector Classifier model achieved the highest performance, with F1 scores of 0.98 for fraud and non-fraud classes. Other models like the Variational Quantum Classifier, Estimator Quantum Neural Network (QNN), and Sampler QNN demonstrate promising results, propelling the potential of QML classification for financial applications. While they exhibit certain limitations, the insights attained pave the way for future enhancements and optimisation strategies. However, challenges exist, including the need for more efficient Quantum algo..

Computational Economics

Is rapid recovery always the best recovery? - Developing a machine learning approach for optimal assignment rules under capacity constraints for knee replacement patients

Recent research suggests that rapid recovery after knee replacement is beneficial for all patients. Rapid recovery requires timely attention after surgery, yet staff resources are usually limited. Thus, patients with the highest possible health gains from rapid recovery should be identified with the objective to prioritise these patients when assigning rapid recovery capacities. We analyze the effect of optimal assignment rules under different capacity constraints for patients set on the rapid recovery care path using disease specific patient-reported outcomes (KOOS-PS) as measure for effectiveness. Subsequently, we build a policy tree to develop optimal treatment assignment rules. We use pa..

Computational Economics

ChatGPT-based Investment Portfolio Selection

In this paper, we explore potential uses of generative AI models, such as ChatGPT, for investment portfolio selection. Trusting investment advice from Generative Pre-Trained Transformer (GPT) models is a challenge due to model "hallucinations", necessitating careful verification and validation of the output. Therefore, we take an alternative approach. We use ChatGPT to obtain a universe of stocks from S&P500 market index that are potentially attractive for investing. Subsequently, we compared various portfolio optimization strategies that utilized this AI-generated trading universe, evaluating those against quantitative portfolio optimization models as well as comparing to some of the popula..

Computational Economics

Should we trust web-scraped data?

The increasing adoption of econometric and machine-learning approaches by empirical researchers has led to a widespread use of one data collection method: web scraping. Web scraping refers to the use of automated computer programs to access websites and download their content. The key argument of this paper is that na\"ive web scraping procedures can lead to sampling bias in the collected data. This article describes three sources of sampling bias in web-scraped data. More specifically, sampling bias emerges from web content being volatile (i.e., being subject to change), personalized (i.e., presented in response to request characteristics), and unindexed (i.e., abundance of a population reg..

Computational Economics

Company Similarity using Large Language Models

Identifying companies with similar profiles is a core task in finance with a wide range of applications in portfolio construction, asset pricing and risk attribution. When a rigorous definition of similarity is lacking, financial analysts usually resort to 'traditional' industry classifications such as Global Industry Classification System (GICS) which assign a unique category to each company at different levels of granularity. Due to their discrete nature, though, GICS classifications do not allow for ranking companies in terms of similarity. In this paper, we explore the ability of pre-trained and finetuned large language models (LLMs) to learn company embeddings based on the business desc..

Computational Economics

IMM: An Imitative Reinforcement Learning Approach with Predictive Representation Learning for Automatic Market Making

Market making (MM) has attracted significant attention in financial trading owing to its essential function in ensuring market liquidity. With strong capabilities in sequential decision-making, Reinforcement Learning (RL) technology has achieved remarkable success in quantitative trading. Nonetheless, most existing RL-based MM methods focus on optimizing single-price level strategies which fail at frequent order cancellations and loss of queue priority. Strategies involving multiple price levels align better with actual trading scenarios. However, given the complexity that multi-price level strategies involves a comprehensive trading action space, the challenge of effectively training profit..

Computational Economics

Variations on the Reinforcement Learning performance of Blackjack

Blackjack or "21" is a popular card-based game of chance and skill. The objective of the game is to win by obtaining a hand total higher than the dealer's without exceeding 21. The ideal blackjack strategy will maximize financial return in the long run while avoiding gambler's ruin. The stochastic environment and inherent reward structure of blackjack presents an appealing problem to better understand reinforcement learning agents in the presence of environment variations. Here we consider a q-learning solution for optimal play and investigate the rate of learning convergence of the algorithm as a function of deck size. A blackjack simulator allowing for universal blackjack rules is also imp..

Computational Economics

Anomaly Detection in Global Financial Markets with Graph Neural Networks and Nonextensive Entropy

Anomaly detection is a challenging task, particularly in systems with many variables. Anomalies are outliers that statistically differ from the analyzed data and can arise from rare events, malfunctions, or system misuse. This study investigated the ability to detect anomalies in global financial markets through Graph Neural Networks (GNN) considering an uncertainty scenario measured by a nonextensive entropy. The main findings show that the complex structure of highly correlated assets decreases in a crisis, and the number of anomalies is statistically different for nonextensive entropy parameters considering before, during, and after crisis.

Computational Economics

Machine Forecast Disagreement

We propose a statistical model of differences in beliefs in which heterogeneous investors are represented as different machine learning model specifications. Each investor forms return forecasts from their own specific model using data inputs that are available to all investors. We measure disagreement as dispersion in forecasts across investor-models. Our measure aligns with extant measures of disagreement (e.g., analyst forecast dispersion), but is a significantly stronger predictor of future returns. We document a large, significant, and highly robust negative cross-sectional relation between belief disagreement and future returns. A decile spread portfolio that is short stocks with high ..

Computational Economics

Emerging Frontiers: Exploring the Impact of Generative AI Platforms on University Quantitative Finance Examinations

This study evaluated three Artificial Intelligence (AI) large language model (LLM) enabled platforms - ChatGPT, BARD, and Bing AI - to answer an undergraduate finance exam with 20 quantitative questions across various difficulty levels. ChatGPT scored 30 percent, outperforming Bing AI, which scored 20 percent, while Bard lagged behind with a score of 15 percent. These models faced common challenges, such as inaccurate computations and formula selection. While they are currently insufficient for helping students pass the finance exam, they serve as valuable tools for dedicated learners. Future advancements are expected to overcome these limitations, allowing for improved formula selection and..

Computational Economics

Artificial Intelligence and Scientific Discovery: A Model of Prioritized Search

We model a key step in the innovation process, hypothesis generation, as the making of predictions over a vast combinatorial space. Traditionally, scientists and innovators use theory or intuition to guide their search. Increasingly, however, they use artificial intelligence (AI) instead. We model innovation as resulting from sequential search over a combinatorial design space, where the prioritization of costly tests is achieved using a predictive model. We represent the ranked output of the predictive model in the form of a hazard function. We then use discrete survival analysis to obtain the main innovation outcomes of interest – the probability of innovation, expected search duration, ..

Computational Economics

A New Approach to Overcoming Zero Trade in Gravity Models to Avoid Indefinite Values in Linear Logarithmic Equations and Parameter Verification Using Machine Learning

The presence of a high number of zero flow trades continues to provide a challenge in identifying gravity parameters to explain international trade using the gravity model. Linear regression with a logarithmic linear equation encounters an indefinite value on the logarithmic trade. Although several approaches to solving this problem have been proposed, the majority of them are no longer based on linear regression, making the process of finding solutions more complex. In this work, we suggest a two-step technique for determining the gravity parameters: first, perform linear regression locally to establish a dummy value to substitute trade flow zero, and then estimating the gravity parameters...

Computational Economics

Inventor Gender and Patent Undercitation: Evidence from Causal Text Estimation

Implementing a state-of-the-art machine learning technique for causal identification from text data (C-TEXT), we document that patents authored by female inventors are under-cited relative to those authored by males. Relative to what the same patent would be predicted to receive had the lead inventor instead been male, patents with a female lead inventor receive 10% fewer citations. Patents with male lead inventors tend to undercite past patents with female lead inventors, while patent examiners of both genders appear to be more even-handed in the citations they add to patent applications. For female inventors, market-based measures of patent value load significantly on the citation counts t..

Computational Economics

AI-Assisted Investigation of On-Chain Parameters: Risky Cryptocurrencies and Price Factors

Cryptocurrencies have become a popular and widely researched topic of interest in recent years for investors and scholars. In order to make informed investment decisions, it is essential to comprehend the factors that impact cryptocurrency prices and to identify risky cryptocurrencies. This paper focuses on analyzing historical data and using artificial intelligence algorithms on on-chain parameters to identify the factors affecting a cryptocurrency's price and to find risky cryptocurrencies. We conducted an analysis of historical cryptocurrencies' on-chain data and measured the correlation between the price and other parameters. In addition, we used clustering and classification in order to..

Computational Economics

Nested Multilevel Monte Carlo with Biased and Antithetic Sampling

We consider the problem of estimating a nested structure of two expectations taking the form $U_0 = E[\max\{U_1(Y), \pi(Y)\}]$, where $U_1(Y) = E[X\ |\ Y]$. Terms of this form arise in financial risk estimation and option pricing. When $U_1(Y)$ requires approximation, but exact samples of $X$ and $Y$ are available, an antithetic multilevel Monte Carlo (MLMC) approach has been well-studied in the literature. Under general conditions, the antithetic MLMC estimator obtains a root mean squared error $\varepsilon$ with order $\varepsilon^{-2}$ cost. If, additionally, $X$ and $Y$ require approximate sampling, careful balancing of the various aspects of approximation is required to avoid a signific..

Computational Economics