The Economic Lives of the Poor is a 2007 paper by 2019 Nobel Laureates1Abhijit Banerjee and Esther Duflo. It describes results from household surveys of the poor (less than $2.16 a day in 1993 PPP dollars) and extremely poor (less than $1.08 a day in 1993 PPP dollars) conducted in Cote d’Ivoire, Guatemala, India, Indonesia, Mexico, Nicaragua, Pakistan, Panama, Papua New Guinea, Peru, South Africa, Tanzania, and Timor Leste (East Timor). What follows is a condensation of that paper highlighting the most important and striking claims. I focus on the descriptive part of the paper and omit the second part of the paper investigating some concomitant economic puzzles.
Living arrangements
Extended families are common among the extremely poor with the median household containing between seven and eight members. The population tends to be young with only about one quarter as many over 51s as 21–50s (In the U.S., the ratio is 0.6 rather than ~0.25.).
Spending
Food and other consumption purchases
Food expenditures typically represent between one half and three quarters of the budget.2 Expenditure on alcohol and tobacco ranges from about 1 percent to about 8 percent. The median extremely poor household in Udaipur spent 10 percent of its budget on festivals, but expenditure on movies, theater and shows is less than 1 percent.
Ownership of assets
Land ownership among the extremely poor ranges from 4 percent in Mexico to 99 percent in Udaipur. However, the owned plots are often small—less than three hectares—and of poor quality.
“In Udaipur, where we have detailed asset data, most extremely poor households have a bed or a cot, but only about 10 percent have a chair or a stool and 5 percent have a table. […] No one has a phone.”
Health and well-being
The bottom decile of income in an Indian survey sample averages 1400 calories a day. Sixty-five percent of poor adults in Udaipur have a BMI that classifies them as underweight and 55 percent are anemic.
Between 11 and 46 percent (depending on the country) of the extremely poor report having been bedridden for at least one day in the the last month.
“While the poor certainly feel poor, their levels of self-reported happiness or self-reported health levels are not particularly low (Banerjee, Duflo, and Deaton, 2004). On the other hand, the poor do report being under a great deal of stress, both financial and psychological.”
Much more interesting IMO than these short excerpts will suggest:
For about 20 years, the team led by Vladimir Braginsky at Moscow State University, as part of a larger programmeon low dissipation systems, has been claiming to have measured quality factors (Qs) in sapphire up to 4x10^8 at room temperature. The ‘quality factor’ of a material indicates the rate of decay of its resonances — how long it will ‘ring’ if struck. […] But until the summerof 1999, no one outside Moscow State had succeeded in measuring a Q in sapphire higher than about 5x10’.
In the summer of 1998, after a series of failed efforts to measure Qs comparable to the Russian claims, members of a Glasgow University group visited Moscow State University for a week to learn the Russian technique. […] In neither case was a high-Q measurement achieved. Nevertheless, after only a few days in Russia, the Glasgow team had become convinced that the Russian results were correct.
Studies that deliberately infect people with diseases are on the rise. They promise speedier vaccine development, but there’s a need to shore up informed consent.
Thus, the moral biases of slavery advocates proved largely immune to correction by the dominant methods of moral philosophy, which were deployed by white abolitionists. Ascent to the a priori led to abstract moral principles—the Golden Rule, the equality of humans before God—that settled nothing because their application to this world was contested. Table-turning exercises were ineffective for similar reasons. Reflective equilibrium did not clearly favor the abolitionists, given authoritarian, Biblical, and racist premises shared by white abolitionists and slavery advocates. No wonder only a handful of Southern whites turned against slavery on the basis of pure moral argument.
The donor community has been increasingly concerned that development assistance intended for crucial social and economic sectors might be used directly or indirectly to fund unproductive military and other expenditures. The link between foreign aid and public spending is not straightforward because some aid may be “fungible.” This article empirically examines the impact of foreign aid on the recipient’s public expenditures, using cross-country samples of annual observations for 1971-90
The most plausible candidate [to making linear programming solutions feasible for economic planning problems] is to look for problems which are “separable”, where the constraints create very few connections among the variables. If we could divide the variables into two sets which had nothing at all to do with each other, then we could solve each sub-problem separately, at tremendous savings in time. The supra-linear, \(n^{3.5}\) scaling would apply only within each sub-problem. We could get the optimal prices (or optimal plans) just by concatenating the solutions to sub-problems, with no extra work on our part.
Unfortunately, as Lenin is supposed to have said, “everything is connected to everything else”. […] A national economy simply does not break up into so many separate, non-communicating spheres which could be optimized independently.
So long as we are thinking like computer programmers, however, we might try a desperately crude hack, and just ignore all kinds of interdependencies between variables. If we did that, if we pretended that the over-all high-dimensional economic planning problem could be split into many separate low-dimensional problems, then we could speed things up immensely, by exploiting parallelism or distributed processing. […]
At this point, each processor is something very much like a firm, with a scope dictated by information-processing power, and the mis-matches introduced by their ignoring each other in their own optimization is something very much like “the anarchy of the market”.
(This post is painfully long. Coping advice: Each subsection within Direct (empirical) evidence, within Indirect evidence, and within Responses is pretty independent—feel free to dip in and out as desired. I’ve also put a list-formatted summary at the end of each these sections boiling down each subsection to one or two sentences.)
Intro
Dan is a student council representative at his school. This semester he is in charge of scheduling discussions about academic issues. He often picks topics that appeal to both professors and students in order to stimulate discussion.
Is Dan’s behavior morally acceptable? On first glance, you’d be inclined to say yes. And even on the second and third glance, obviously, yes. Dan is a stand-up guy. But what if you’d been experimentally manipulated to feel disgust while reading the vignette? If we’re to believe (Wheatley and Haidt 2005), there’s a one-third chance you’d judge Dan as morally suspect. ‘One subject justified his condemnation of Dan by writing “it just seems like he’s up to something.” Another wrote that Dan seemed like a “popularity seeking snob.”’
The possibility that moral judgments track irrelevant factors like incidental disgust at the moment of evaluation is (to me, at least) alarming. But now that you’ve been baited, we can move on the boring, obligatory formalities.
Arguably, we don’t care about the exact cost-effectiveness estimates of each of GiveWell’s top charities. Instead, we care about their relative values. By using distance metrics across these multidimensional outputs, we can perform uncertainty and sensitivity analysis to answer questions about:
how uncertain we are about the overall relative values of the charities
which input parameters this overall relative valuation is most sensitive to
In the lasttwo posts, we performed uncertainty and sensitivity analyses on GiveWell’s charity cost-effectiveness estimates. Our outputs were, respectively:
probability distributions describing our uncertainty about the value per dollar obtained for each charity and
estimates of how sensitive each charity’s cost-effectiveness is to each of its input parameters
Another issue is that by treating each cost-effectiveness estimate as independent we underweight parameters which are shared across many models. For example, the moral weight that ought to be assigned to increasing consumption shows up in many models. If we consider all the charity-specific models together, this input seems to become more important.
Metrics on rankings
We can solve both of these problems by abstracting away from particular values in the cost-effectiveness analysis and looking at the overall rankings returned. That is we want to transform:
GiveWell’s cost-effectiveness estimates for its top charities
Charity
Value per $10,000 donated
GiveDirectly
38
The END Fund
222
Deworm the World
738
Schistosomiasis Control Initiative
378
Sightsavers
394
Malaria Consortium
326
Against Malaria Foundation
247
Helen Keller International
223
into:
Givewell’s top charities ranked from most cost-effective to least
Deworm the World
Sightsavers
Schistosomiasis Control Initiative
Malaria Consortium
Against Malaria Foundation
Helen Keller International
The END Fund
GiveDirectly
But how do we usefully express probabilities over rankings1 (rather than probabilities over simple cost-effectivness numbers)? The approach we’ll follow below is to characterize a ranking produced by a run of the model by computing its distance from the reference ranking listed above (i.e. GiveWell’s current best estimate). Our output probability distribution will then express how far we expect to be from the reference ranking—how much we might learn about the ranking with more information on the inputs. For example, if the distribution is narrow and near 0, that means our uncertain input parameters mostly produce results similar to the reference ranking. If the distribution is wide and far from 0, that means our uncertain input parameters produce results that are highly uncertain and not necessarily similar to the reference ranking.
Spearman’s footrule
What is this mysterious distance metric between rankings that enables the above approach? One such metric is called Spearman’s footrule distance. It’s defined as:
\(c\) varies over all the elements \(A\) of the rankings and
\(\text{pos}(r, x)\) returns the integer position of item \(x\) in ranking \(r\).
In other words, the footrule distance between two rankings is the sum over all items of the (absolute) difference in positions for each item. (We also add a normalization factor so that the distance varies ranges from 0 to 1 but omit that trivia here.)
So the distance between A, B, C and A, B, C is 0; the (unnormalized) distance between A, B, C and C, B, A is 4; and the (unnormalized) distance between A, B, C and B, A, C is 2.
Kendall’s tau
Another common distance metric between rankings is Kendall’s tau. It’s defined as:
\(i\) and \(j\) are items in the set of unordered pairs \(P\) of distinct elements in \(u\) and \(v\)
\(\bar{K}_{i,j}(u, v) = 0\) if \(i\) and \(j\) are in the same order (concordant) in \(u\) and \(v\) and \(\bar{K}_{i,j}(u, v) = 1\) otherwise (discordant)
In other words, the Kendall tau distance looks at all possible pairs across items in the rankings and counts up the ones where the two rankings disagree on the ordering of these items. (There’s also a normalization factor that we’ve again omitted so that the distance ranges from 0 to 1.)
So the distance between A, B, C and A, B, C is 0; the (unnormalized) distance between A, B, C and C, B, A is 3; and the (unnormalized) distance between A, B, C and B, A, C is 1.
Angular distance
One drawback of the above metrics is that they throw away information in going from the table with cost-effectiveness estimates to a simple ranking. What would be ideal is to keep that information and find some other distance metric that still emphasizes the relationship between the various numbers rather than their precise values.
Angular distance is a metric which satisfies these criteria. We can regard the table of charities and cost-effectiveness values as an 8-dimensional vector. When our output produces another vector of cost-effectiveness estimates (one for each charity), we can compare this to our reference vector by finding the angle between the two2.
Visual (scatter plot) and delta moment-independent sensitivity analysis on GiveWell’s cost-effectiveness models show which input parameters the cost-effectiveness estimates are most sensitive to. Preliminary results (given our input uncertainty) show that some input parameters are much more influential on the final cost-effectiveness estimates for each charity than others.
Last time we introduced GiveWell’s cost-effectiveness analysis which uses a spreadsheet model to take point estimates of uncertain input parameters to point estimates of uncertain results. We adjusted this approach to take probability distributions on the input parameters and in exchange got probability distributions on the resulting cost-effectiveness estimates. But this machinery lets us do more. Now that we’ve completed an uncertainty analysis, we can move on to sensitivity analysis.
Sensitivity analysis
The basic idea of sensitivity analysis is, when working with uncertain values, to see which input values most affect the output when they vary. For example, if you have the equation \(f(a, b) = 2^a + b\) and each of \(a\) and \(b\) varies uniformly over the range from 5 to 10, \(f(a, b)\) is much more sensitive to \(a\) then \(b\). A sensitivity analysis is practically useful in that it can offer you guidance as to which parameters in your model it would be most useful to investigate further (i.e. to narrow their uncertainty).
Visual sensitivity analysis
The first kind of sensitivity analysis we’ll run is just to look at scatter plots comparing each input parameter to the final cost-effectiveness estimates. We can imagine these scatter plots as the result of running the following procedure many times1: sample a single value from the probability distribution for each input parameter and run the calculation on these values to determine a result value. If we repeat this procedure enough times, it starts to approximate the true values of the probability distributions.
(One nice feature of this sort of analysis is that we see how the output depends on a particular input even in the face of variations in all the other inputs—we don’t hold everything else constant. In other words, this is a global sensitivity analysis.)
(Caveat: We are again pretending that we are equally uncertain about each input parameter and the results reflect this limitation. To see the analysis result for different input uncertainties, edit and run the Jupyter notebook.)
Direct cash transfers
GiveDirectly
Scatter plots showing sensitivity of GiveDirectly’s cost-effectiveness to each input parameter
The scatter plots show that, given our choice of input uncertainty, the output is most sensitive (i.e. the scatter plot for these parameters shows the greatest directionality) to the input parameters:
Highlighted input factors to which result is highly sensitive
Input
Type of uncertainty
Meaning/importance
value of increasing ln consumption per capita per annum
Moral
Determines final conversion between empirical outcomes and value
transfer as percent of total cost
Operational
Determines cost of results
return on investment
Opportunities available to recipients
Determines stream of consumption over time
baseline consumption per capita
Empirical
Diminishing marginal returns to consumption mean that baseline consumption matters