This literature review focuses on the contribution of the heterogeneous agents framework to the empirical robustness of macroeconomic models.
First, we focus on the transmission of monetary policy in an economy characterized by heterogeneous agents. We do this by analyzing both the quantitative HANK model and an analytical representation (THANK).
Secondly, we illustrate the greater suitability of heterogeneous agent models for economic and welfare analysis in a developing country context.
Finally, we analyze agent-based models as a potential avenue to address a higher degree of heterogeneity and complexity in the data.
HANK and THANK models
We see that the HANK model delivers a more accurate representation of the wealth and consumption distribution of households, but it still lacks important dimensions of household heterogeneity. For instance, the distribution of capital gains is crucial to match the empirical evidence on movements of capital and equity prices.
The THANK model attempts to give a tractable representation of the HANK model, keeping a certain degree of idiosyncratic uncertainty. Yet, in contrast with the latter, it is not able to address issues related to wealth distribution and welfare.
The challenges of working with macroeconomic simulations
Even with heterogeneous agents, it is unclear whether macroeconomic simulations of the scaling up of micro-evaluations can really do more than making us aware and cautious of forces in general equilibrium that might alter the effects found in RCTs.
Given the data-challenges in developing countries, the dimension of the informal economy, and in general the immense complexity that remains beyond what is captured in the models, the model predictions might still be inaccurate.
The introduction of agent-based models attempts to represent more complex and richer economic dynamics. However, this enhancement comes with some drawbacks. First, researchers are left with almost arbitrary freedom in choosing the inputs of the models (e.g., behavioral equations governing agents’ behavior). Second, causal mechanisms in the model are unclear (“black box” critique).
Better models for complex, real-world economies
Despite these challenges, the introduction of the heterogeneous agents’ framework makes macroeconomic models more suitable to capture the complex dynamics in real-world economies.
We conclude that such development in the macroeconomic field is key to enhance the relevance of the models’ policy prescriptions and to improve the ability of their empirical performance.
Stacking is an ensemble method that is used widely in supervised learning. As always, we have some training data, and the goal is to predict target variable in new data. The basic structure of the method consists in two levels of learners: base and meta. The main idea is that the meta learner combines predictions of the base learners to provide final response. That is, predictions of the base learners are used to calibrate the meta learner. The beauty of ensemble techniques in general is that they allow to capture various aspects of patterns in data using heterogeneous machine learning models. It explains why ensembles often happen to exhibit strong predictive power. For peak performance, it is recommended to have some degree of diversity in the base learners and sufficiently small correlation among their predictions.
Of course, at first sight, this algorithm might seem to be prone to overfitting. However, stacking is constructed in a way that helps to avoid it. In fact, it is the most subtle part of the technique. In short, while calibrating the meta learner, one uses cross-validation type of procedure to combine predictions of the base learners.
Along the text, I will use some pseudo-code notation which I will be defining on the go. Also, the text was prepared having in mind the stacking ensemble implementations from scikit-learn and mlens Python packages.
The purpose of this text is to describe meta learner calibration and base learners training separately to avoid any misinterpretations that could arise when mixing these components together.
I start with a short description of a basic supervised learning setup to define some notation. Then, I proceed to how the meta learner is trained. Finally, I explain how one obtains predictions for new data. Take a coffee, and let’s jump in!
Setup
In this text, we consider a basic numerical data setup. That is, we have a target which we want to predict, and the feature matrix which carries information about some features of n observations.
Let’s quickly setup some notation:
Data: X_train (feature matrix), y_train (target).
Number of rows (observations) in the training data: n.
Cross-validation folds (disjoint): cv_folds = [fold_1, …, fold_K], such that union(cv_folds) = {1, … , n}.
Base learners: base_learners = [base_learner[1], ..., base_learner[B]].
Meta learner: meta_learner.
New data: X_new.
The learners here are generic objects that define a learning model, that can be fitted on data via applying a generic method .fit() and predict a target via method .predict(). Whenever the learners are trained, I mention it explicitly. With this setup, we proceed to describing how one trains the meta learner.
Stacking Ensemble Training
Now, we are ready to discuss how we can train a stacking ensemble and then obtain predictions for new data.
Meta learner calibration
One first fits the base learners on the training data using a cross-validation approach. Remember, our final goal here is to train the meta learner, which will be producing final target predictions. So, we do not want to propagate information about target realisation of the i-th observation into a base learner prediction for it since doing this will cause extreme overfitting of the meta learner. That is why we use a cross-validation type of procedure. Then, when we have predictions of the base learners obtained via the cross-validation approach, we can concatenate them with our initial feature matrix, i.e. considering them as new features for the meta learner. So, to train the meta learner one simply fits it to the target given the extended feature matrix. That is it!
Now, let’s summarise the procedure into a piece of pseudo-code:
Note that construction of X_meta may vary from one implementation to another. For example, one may choose a subset of features to use (propagate) in the meta-learner. Attention: do not confuse it with a concept of feature propagation that was introduced to cope with a problem of missing data in learning on graphs by Rossi, E., Kenlay, H., Gorinova, M.I., Chamberlain, B.P., Dong, X. and Bronstein, M.M. in the paper “On the unreasonable effectiveness of feature propagation in learning on graphs with missing node features.” Learning on Graphs Conference; PMLR, 2022.
As we can see, the structure is not complex. Of course, there are some technicalities that do not appear in this demonstrative pseudo-code. For example, one might want to do the cross-validation and base learners training in parallel to speed up performance. For a reference on the state-of-the-art implementation in Python, check out scikit-learn and mlens packages.
Meta learner training
Base learners training
I describe this step after meta learner calibration on purpose so that we do not confuse the former with a part of the latter. So far, we have calibrated the meta learner. However, to predict the target for new data we also need to train the base learners so they are able to produce base predictors for the unseen dataset.
The procedure is straightforward. One can just fit each base learner on the whole training dataset X_train, y_train. Again, when training the meta learner we were using the cross-validation approach to avoid excessive overfitting of the meta learner. Here, we do not have to do this since we will use the base learners to obtain base predictions for newly occurred data that was never seen by the model. So, we can utilize all the data that we have.
Stacking Ensemble prediction
At this point, we have the meta learner calibrated and the base learners trained. So, to predict target values for newly occurred data, we first obtain the base target predictions, and then use the meta learner.
Conclusion
Ensemble methods are particularly known for their decent prediction performance in supervised-learning setups if used appropriately. Stacking ensembles exhibit hierarchical structure with two levels: base and meta. Meta learner combines responses of base learners to provide final prediction of a target variable. To avoid overfitting, meta learner is trained involving cross-validation type of procedure used to obtain base-learner predictions.
I discuss how stacking ensembles can be trained and used for prediction in supervised learning problems. I decided to maintain some level of generality in the method description while providing pseudo-code examples.
The rise in global temperatures has been a growing source of concern for policy makers given its potential impact for sustainable economic activity.
This paper explores the effects of a temperature shock on the distribution of economic growth in Colombia. Specifically, we focus on the impact of these shocks on economic risks.
Our findings suggest the existence of asymmetries on the lower and upper tail of the distribution, indicating that higher temperatures leads to less optimistic outcomes for GDP growth. We also found negative and significant effects of temperature shocks on non-agricultural sectors.
The results are obtained using quantile regressions under the growth-at-risk (GaR) approach and local projections.
Temperature shocks on climate-exposed sectors
The analysis of the effects of temperature shocks on climate-exposed sectors suggest that several components of GDP are affected by increased heating.
Manufacturing activity seems to be negatively affected after a temperature shock, specially on the upper tail.
The entire distribution of construction shows a strong response to temperature shocks, suggesting that labor productivity could be severely affected by heat stress in this sector.
The generation of electricity by hydro-powers in Colombia could be explaining the negative effects on the energy sector after an increase in temperatures.
Results for agriculture are not statistically significant, possibly capturing the national attempt to increase resilience of this sector in Colombia.
Protecting Colombia’s economy from the impacts of global warming
Colombia has taken some steps in the mitigation and adaptation of its economy to rising temperatures. Especially, it has adopted the National Adaptation Program, whose principal objective is to reduce the country’s vulnerability and increase its response capacity to the impacts of global warming on economic growth, especially for the agricultural sector.
The findings for this sector and the mitigating measures adopted suggest that these efforts could reduce the country’s vulnerability to rising temperatures. However, the country should take complementary strategies with a broader scope of sectors such as energy and construction, that appear to be more vulnerable to warming.
For example, in the energy sector some of 18 the adaptation measures that could be taken are optimization of the conventional energies and improvement of efficiency and diversification of energy sources and promotion of renewal energy.
Further research
Further analysis about the channels and causes of the effects of temperature shocks on GDP growth could be done for a regional level. The heterogeneity on temperature levels and on the reliance on climate-exposed sectors across different regions in Colombia could imply differential effects depending the geographical area analyzed.
In this literature review, we focus on workplace discrimination with particular attention paid to the hiring process. We classified the hiring process into two stages: CV-screening and interview.
Following the theoretical discussion of stereotypes supporting the existence of discrimination, two empirical pieces of research are introduced:
The first paper shows that race affects applicants’ callback rate during the screening process
The second study shows that the probability of a female musician being hired increases with the introduction of blind auditions.
Lastly, from a sociological aspect, we claim that cultural similarities can affect employers’ hiring decisions, from which workplace discrimination might also appear.
All four of the authors are currently PhD students in Economics in the doctoral program organized jointly by Universitat Pompeu Fabra and the Barcelona School of Economics. (March 2023)
This paper examines the downgrading in job status that immigrant workers suffer when settling in a new country. We consider the massive Venezuelan exodus and the impacts this shock had on the job outcomes of migrants who settled in Chile.
Our approach is based on linear regression analysis and multinomial logistic regression models to estimate the penalty immigrants face. To this end, we use household-level data and employ two job-status indexes.
Results show that migrants who arrived before the 2015 Venezuelan crisis did not face significant downgrading. However, migrants who have arrived after 2015 do. Findings are relevant to understanding the impact of massive and sudden migratory shocks.
In this paper we consider the negative medium-term impact of the Russian Financial Crisis of 2014-2016 on remittances, consumption patterns, and credit in the Kyrgyz Republic.
Using panel data from the Life in Kyrgyzstan survey, we show a significant drop in migration to Russia and remittances from Russia on the extensive and intensive margins.
Households with a migrant abroad in Russia just prior to the crisis experience an average fall in real per capita income and an increase in poverty, while households both with and without a migrant rebalance their consumption basket to cope with the economic downturn.
Economics of Public Policy alum Sam Juthani ’13 recently wrote about the maturing digital assets market in a blog post for Flint Global, where he works in the Markets and Investor Advisory practice.
BSE Economics of Public Policy alum Sam Juthani ’13 recently wrote about the maturing digital assets market in a blog post for Flint Global, where he works in the Markets and Investor Advisory practice.
Sam shared an overview of his post on LinkedIn:
“I’ve blogged about the outlook for crypto and DLT (distributed ledger technology). While crypto assets might have hogged the headlines, the big story is the investment in the underlying infrastructure. And that’s where there’s a huge amount of commercial possibility – from payments, digital bonds, tokenisation, and safe, stable ways to access a wider digital economy.”
He pointed out that “2023 is a major year for digital regulation, and there’s a risk that innovation is stopped in its tracks by policymakers who are rightly trying to stop cases of abuse and fraud. This isn’t a question of balance so much as understanding – and businesses have an important role to play in helping policymakers understand that world.”
Publication in Nature Medicine by Roger Garriga ’17 and Javier Mas ’17 (Data Science) et al
The use of machine learning in healthcare is still in its infancy. In this paper, we describe the project we did to predict psychotic episodes together with Birmingham’s psychiatric hospital. We hope to see these sorts of applications of ML in healthcare become the new standard in the future. The technology is ready, so it’s just a matter of getting it done!
Paper abstract
The timely identification of patients who are at risk of a mental health crisis can lead to improved outcomes and to the mitigation of burdens and costs. However, the high prevalence of mental health problems means that the manual review of complex patient records to make proactive care decisions is not feasible in practice. Therefore, we developed a machine learning model that uses electronic health records to continuously monitor patients for risk of a mental health crisis over a period of 28 days. The model achieves an area under the receiver operating characteristic curve of 0.797 and an area under the precision-recall curve of 0.159, predicting crises with a sensitivity of 58% at a specificity of 85%. A follow-up 6-month prospective study evaluated our algorithm’s use in clinical practice and observed predictions to be clinically valuable in terms of either managing caseloads or mitigating the risk of crisis in 64% of cases. To our knowledge, this study is the first to continuously predict the risk of a wide range of mental health crises and to explore the added value of such predictions in clinical practice.
Garriga, R., Mas, J., Abraha, S. et al. Machine learning model to predict mental health crises from electronic health records. Nat Med (2022). https://doi.org/10.1038/s41591-022-01811-5
Publication in Energy Economics by Kadir Özen ’21 (PhD Track) and Dilem Yıldırım
This paper has been published in the November 2021 issue of the journal Energy Economics.
Abstract
The electricity price forecasting (EPF) is a challenging task not only because of the uncommon characteristics of electricity but also because of the existence of many potential predictors with changing predictive abilities over time. In such an environment, how to account for all available factors and extract as much information as possible is the key to the production of accurate forecasts. To address this long-standing issue in a way that balances complexity and forecasting accuracy while facilitating the traceability of the predictor selection procedure, we propose the method of Bootstrap Aggregation (bagging). To forecast day-ahead electricity prices in a multivariate context for six major power markets, we construct a large-scale pure price model and apply the bagging approach in comparison with the popular Least Absolute Shrinkage and Selection Operator (LASSO) estimation method. Our forecasting study reveals that bagging provides substantial forecast improvements on daily and hourly scales in almost all markets over the popular LASSO estimation method. The differentiation in the forecast performances of the two approaches appears to arise, inter alia, from their structural differences in the explanatory variables selection process. Moreover, to account for the intraday hourly dependencies of day-ahead electricity prices, all our models are augmented with latent factors, and a substantial improvement is observed only in the forecasts from models covering a relatively limited number of predictors.
Highlights
We forecast day-ahead electricity prices for major markets with a large-scale model.
The method of Bootstrap Aggregation (bagging) is applied to generate forecasts.
Bagging appears to be very competitive and promising compared to the popular LASSO.
Factor augmentation is proposed to capture intraday hourly dependencies of prices.
Augmentation improves forecasts only for models with limited number of predictors.
Their paper, “Female entrepreneurship, financial frictions and capital misallocation in the US,” has also been published in the Journal of Monetary Economics.
EEA Young Economist Awardees Marta Morazzoni ’18 and Andrea Sy ’18
BSE alumni Marta Morazzoni and Andrea Sy (both Economics Class of 2018) received the 2021 Young Economist Award from the European Economic Association and Unicredit Foundation for their paper, “Female entrepreneurship, financial frictions and capital misallocation in the US.”
The EEA Young Economics award committee consisted of Philipp Kircher, Giacomo Ponzetto and Antonella Trigari. They noted that “the paper addresses an extremely important topic, offers new empirical evidence from micro-level data cleverly identifying informative moments, and builds a state-of-the-art general equilibrium model to rationalize the evidence and to provide highly relevant policy implications.”
Humbled and honored to win the Young Economist Award from @EEANews along with other incredible papers!!! We are grateful to all the great feedback and to the professors and colleagues that helped our work (and us!) grow in this past year https://t.co/QUIOjXM7A6
We document and quantify the effect of a gender gap in credit access on both entrepreneurship and input misallocation in the US. Female entrepreneurs are found to be more likely to face a rejection on their loan applications and to have a higher average product of capital, a sign of gender-driven capital misallocation that decreases in female-led firms’ access to finance. These results are not driven by differences in observable individual or businesses characteristics. Calibrating a heterogeneous agents model of entrepreneurship to the US economy, we show that the observed gap in credit access explains the bulk of the gender differences in capital allocation across firms. Eliminating such credit imbalance is estimated to potentially increase output by 4%, and to reduce capital misallocation by 12%.
Key findings
In the US, female entrepreneurs receive less business funding compared to male entrepreneurs.
Female-owned firms operate with lower levels of assets, resulting in gender-driven capital misallocation.
Female-led businesses are nonetheless relatively more profitable and have better credit risk scores.
Removing the gender gap in business financing is estimated to potentially increase output by 4%.
We use cookies to ensure that we offer you the best user experience on our website. If you accept, we understand that you consent to their use. You can always adjust your preferences using the "Cookie Settings" located in the footer.
Functional
Always active
The technical storage or access is strictly necessary for the legitimate purpose of enabling the use of a specific service explicitly requested by the subscriber or user, or for the sole purpose of carrying out the transmission of a communication over an electronic communications network.
Preferences
The technical storage or access is necessary for the legitimate purpose of storing preferences that are not requested by the subscriber or user.
Statistics
The technical storage or access that is used exclusively for statistical purposes.The technical storage or access that is used exclusively for anonymous statistical purposes. Without a subpoena, voluntary compliance on the part of your Internet Service Provider, or additional records from a third party, information stored or retrieved for this purpose alone cannot usually be used to identify you.
Marketing
The technical storage or access is required to create user profiles to send advertising, or to track the user on a website or across several websites for similar marketing purposes.
We use cookies to optimize our website and our service.
Functional
Always active
The technical storage or access is strictly necessary for the legitimate purpose of enabling the use of a specific service explicitly requested by the subscriber or user, or for the sole purpose of carrying out the transmission of a communication over an electronic communications network.
Preferences
The technical storage or access is necessary for the legitimate purpose of storing preferences that are not requested by the subscriber or user.
Statistics
The technical storage or access that is used exclusively for statistical purposes.The technical storage or access that is used exclusively for anonymous statistical purposes. Without a subpoena, voluntary compliance on the part of your Internet Service Provider, or additional records from a third party, information stored or retrieved for this purpose alone cannot usually be used to identify you.
Marketing
The technical storage or access is required to create user profiles to send advertising, or to track the user on a website or across several websites for similar marketing purposes.