Measure is unceasing

Shallow evaluations of longtermist organizations

Epistemic status: Fairly uncertain. May contain errors, probabilities might not be calibrated.

Introduction

This document reviews a number of organizations in the longtermist ecosystem, and poses and answers a number of questions which would have to be answered to arrive at a numerical estimate of their impact. My aim was to see how useful a “quantified evaluation” format in the longtermist domain would be.

In the end, I did not arrive at GiveWell-style numerical estimates of the impact of each organization, which could be used to compare and rank them. To do this, one would have to resolve and quantify the remaining uncertainties for each organization, and then convert each organization’s impact to a common unit [1, 2].

In the absence of fully quantified evaluations, messier kinds of reasoning have to be used and are being used to prioritize among those organizations, and among other opportunities in the longtermist space. But the hope is that reasoning and reflection built on top of quantified predictions might prove more reliable than reasoning and reflection alone.

In practice, the evaluations below are at a fairly early stage, and I would caution against taking them too seriously and using them in real-world decisions as they are. By my own estimation, of two similar past posts, 2018-2019 Long Term Future Fund Grantees: How did they do? had 2 significant mistakes, as well as half a dozen minor mistakes, out of 24 grants, whereas Relative Impact of the First 10 EA Forum Prize Winners had significant errors in at least 3 of the 10 posts it evaluated.

To make the scope of this post more manageable, I mostly did not evaluate organizations included in Lark’s yearly AI Alignment Literature Review and Charity Comparison posts, nor meta-organizations [3].

Evaluated organizations

Alliance to Feed the Earth in Disasters

Epistemic status for this section: Fairly sure about the points related to ALLFED’s model of its own impact. Unsure about the points related to the quality of ALLFED’s work, given that I’m relying on impressions from others.

Questions

With respect to the principled case for an organization to be working on the area:

  1. What is the probability of a (non-AI) catastrophe which makes ALLFED’s work relevant (i.e., which kills 10% or more of humanity, but not all of humanity) over the next 50 to 100 years?
  2. How much does the value of the future diminish in such a catastrophe?
  3. How does this compare to work in other areas?

With respect to the execution details:

  1. Is ALLFED making progress in its “feeding everyone no matter what” agenda?
  2. Is that progress on the lobbying front, or on the research front?
  3. Is ALLFED producing high-quality research? On a Likert scale of 1-5, how strong are their papers and public writing?
  4. Is ALLFED cost-effective?
  5. Given that ALLFED has a large team, is it a positive influence on its team members? How would we expect employees and volunteers to rate their experience with the organization?

Tentative answers

Execution details about ALLFED in particular

Starting from a quick review as a non-expert, I was inclined to defer to ALLFED’s own expertise in this area, i.e., to trust their own evaluation that their own work was of high value, at least compared to other possible directions which could be pursued within their cause area. Per their ALLFED 2020 Highlights, they are researching ways to quickly scale alternative food production, at the lowest cost, in the case of large catastrophes, i.e., foods which could be produced for several years if there was a nuclear war which blotted out the sun.

However, when talking with colleagues and collaborators, some had the impression that ALLFED was not particularly competent, nor its work high quality. I would thus be curious to see an assessment by independent experts about how valuable their work seems in comparison to other work in their area, or to potential work which could be done.

In 2020, ALLFED also did some work related to the COVID-19 pandemic. While there is a case to be made that the pandemic is a kind of test run for a global catastrophe, I feel that this was a bit of a distraction from their core work.

It’s unclear to me whether their research process is particularly cost-efficient; I’ve made inquiries as to the number of full-time employees (FTEs) for 2020 and its budget for that year, but haven’t been answered. The data about ALLFED’s budget was not available on their webpage. Because they are not a 503 registered charity, a Form 990 isn’t anywhere to be found. It is also not clear to me how many FTEs ALLFED is employing, and how many of those are dedicated to research (vs logistical support, bureaucracy, etc.)

The principled case for an organization working in the area

With regards to the chance of catastrophic risks which would make this work valuable, one guide here is Michael Aird’s database of existential risk estimates, another one is Luisa Rodríguez’s work on estimates of the probability of nuclear wars of varying severity. Interestingly, my intuitive estimates vary depending on whether I ask about estimates per year, or estimates in the next 100 years [4].

ALLFED has used this guesstimate model (taken from the post Cost-Effectiveness of Foods for Global Catastrophes: Even Better than Before?) to estimate its own (future) impact. For instance, the ALLFED 2020 Highlights post mentions the following while linking to the model:

I continue to believe that ALLFED’s work offers the highest expected value at the margin for improving the long-term future and saving expected lives in the present generation

The model itself gives:

~60% confidence of greater cost-effectiveness than AI for the 100 millionth dollar, and ~95% confidence of greater cost-effectiveness at the margin now than AI. Anders Sandberg’s version of the model produced ~80% and ~100% confidence, respectively.

The model presents some structure to estimate ALLFED’s impact, namely:

It seems a little bit confusing at first, but it becomes more clear once you go through it cell by cell. In any case, I disagree pretty heavily with some of the estimates in that model, though I appreciate that it’s a quantified model that gives something to disagree about.

Disagreements and Uncertainties

With those inputs, I arrive, per this guesstimate model at a roughly 50% probability that “marginal money now on alternate foods is more cost effective than on AI risk mitigation”. This is in stark contrast with the original 95%, and at a 15% probability that $100M to alternate foods is “more cost-effective than to AI risk mitigation”. I endorse the 50%, but not the 15%; I’d probably be higher on the latter.

I feel that that 50% is still pretty good, but the contrast between it and the model’s initial 95% is pretty noticeable to me, and makes me feel that the 95% is uncalibrated/untrustworthy. On the other hand, my probabilities above can also be seen as a sort of sensitivity analysis, which shows that the case for an organization working on ALLFED’s cause area is somewhat more robust than one might have thought.

Concluding Thoughts

In conclusion, I disagree strongly with ALLFED’s estimates (probability of cost overruns, impact of ALLFED’s work if deployed, etc.), however, I feel that the case for an organization working in this area is relatively solid. My remaining uncertainty is about ALLFED’s ability to execute competently and cost-effectively; independent expert evaluation might resolve most of it.

Sources


All-Party Parliamentary Group for Future Generations (APPGFG)

Epistemic status for this section: Very sure that APPGFG is a very inexpensive opportunity, less sure about other considerations.

Questions:

Tentative answers

General considerations

Per this writeup, the APPGFG

  1. Has been figuring out how best to influence policy in the UK parliament to care more about future generations.
  2. Campaigned for an “UK Future Generations Bill to embed a Commissioner for Future Generations into the structures of UK policy making”, and successfully lobbied the House of Lords to establish a “Special Inquiry Committee on Risk Assessment and Risk Management,” on how the UK prepares for future risks (beyond pandemics) and works internationally to prepare for global risks, which will work for one year.
  3. Has been building relationships with parliamentarians. They grew a parliamentary group to include 75 parliamentarians, which can be found here. APPGFG also organized various activities for that group.
  4. Has been researching possible policy suggestions: diving into policy areas, and “general research into how the effective altruism community should approach policy, risks and measuring the impact of policy interventions.”

Their overall theory of impact (referred to here) seems straightforward and plausible. I would further add a step where successful policy change in the UK could spur further change in other countries, particularly in the European sphere.

I’m not really sure what their network of external advisors looks like; APPGFG’s post mentions receiving positive feedback from the Future of Humanity Institute (FHI), the Center for the Study of Existential Risk (CSER), the UK civil service, and unspecified others. I would be comparatively more excited if the APPGFG’s external advisors mostly come from FHI, rather than CSER, about which I have some reservations (more on which below, in CSER’s own section).

The APPGFG spent roughly $40k for one full-time employee during 2020. This seems very inexpensive. If the APP wanted to expand and thought they had someone they wanted to hire, it would be at the top of my list. It also seems likely that APPGFG’s two existing employees could be paid better.

This APPGFG’s writeup emphasizes that they have “not yet caused any actual changes to UK government policy”, but insofar as what they’re doing is capacity building, I find their capacity building work promising.

My understanding is that right now, there aren’t that many longtermist related proposals which the APPGFG is able to bring forward, and that the longtermist community itself is uncertain about what kinds of policy proposals to push for. To clarify, my understanding is that policy-wise there is some work the APPGFG can do, such as pushing for the aforementioned Future Generations Bill, nudging legislation in a more longtermist direction, or maybe help shape the UK’s attempt at reducing the likelihood of future COVID-19-like catastrophes. However, these proposals seem relatively small in comparison to what a “longtermist policy agenda” could be, and in fact there isn’t an ambitious “longtermist policy agenda” that the APPGFG can just push for.

With that in mind, the APPGFG’s strategy of embedding itself into Britain’s parliamentary processes, while thinking about which more ambitious policy proposals could be brought forward in the future, seems sensible.

Possible downsides

With regards to possible downsides to the APPGFG, the main one in the common EA consciousness seems to be “poisoning the well”. This refers to a possible path whether early suboptimal exposure to longtermist ideas could make the audiences more reticent to later consider similar ideas.

Two other downsides are 1) the APPGFG’s current leadership getting promoted to incompetence in case the APPGFG grows substantially, and 2) the APPGFG’s existence impeding the creation and growth of a more capable organization.

In the first case, maybe the APPGFG’s current leadership are good lobbyists and good researchers, but would be unsuitable to lead e.g., a 20 person lobbying apparatus (and would fail to grow into the position.) But by the time the APPGFG was considering growing that much, it would be awkward to replace its leadership. In the second case, maybe there is a more promising person out there who would have done something similar to the APPGFG, but better, and who didn’t because the APPGFG already existed.

My impression is that this “promotion to incompetence” dynamic may have happened in some EA research organizations, and that the Iodine Global Network may have been both too weak to establish strong, well-funded national groups, and so large that the creation of another organization to do that would be extremely awkward.

In the counterfactual world where the APPGFG didn’t exist, one would still have to develop a policy agenda, and then in addition one would also have to gain familiarity with the British policy-making apparatus, and a presence within it. Whereas in the world where the APPGG does exist, one can develop a longtermist policy agenda informed by political realities, and one has a >2 year head start in establishing a presence in the British political apparatus.

Earlier capacity building seems to me to be worth some poisoning the well, and the overall probability of poisoning the well seems to me to be low. Promotion to incompetence would only be a worry if the APPGFG substantially expanded. Displacing other potentially better organizations seems (to me) to be more of a concern. But overall I think we live in a world where there are not enough people with policy expertise doing EA work, not in the world where there are many and the best are displaced.

Conclusion

In conclusion, I feel that their logic model is solid, and that the APPGFG’s capacity-building work is promising. I’m hesitant about its closeness to CSER. It’s current budget seems particularly small. I’m uncertain about how they compare with other organizations in similar or adjacent spheres, and in particular with GovAI. Downsides exist, but accelerated capacity building seems to me to be worth these downsides.

I feel fairly positive about the APPGFG’s chances of success:

Sources


CSER

Epistemic status for this section: Unmitigated inside view.

Questions

Tentative answer

A colleague mentioned that there was something “weird” with CSER going on, and I was surprised to find out that this was actually the case.

I skimmed the past research of the members mentioned on their webpage, and I classified their researchers in terms of alignment. I came away with the impression that they had around 5 aligned researchers, around 4 researchers I’m uncertain about, and around 14 whom I’d classify as unaligned or unproductive. CSER also has 6 additional support staff.

Readers are welcome to browse CSER’s team page and calculate what percentage of researchers are working on valuable directions according to one’s values.

Personally, although I feel like there is a small group of strong researchers working at CSER, the proportion of researchers working on stuff I don’t particularly care about or which I don’t expect to be particularly valuable according to my values is too high. Commenters pointed out that this assessment is “almost unfairly subjective.”

Sources


Center for Security and Emerging Technology (CSET)

Epistemic status for this section: After doing a shallow dive and reading a portion of CSET’s work , I have some models about their impact, but they are fuzzy and I don’t feel particularly sure about them.

Questions

Tentative answers

CSET’s work can be categorized as:

Analyzing each of them in turn, I looked at past testimonies given by CSET team members to the US Senate and House of Representatives:

Personally, I find the testimonies thoughtful and interesting. They distill complex topics into things which US Members of Congress might understand. However, it is unclear to me to what extent these testimonies actually had an impact on policy.

I thought that testimonies were particularly important because one worry outlined in Open Philanthropy’s grant to found CSET was:

We worry that heavy government involvement in, and especially regulation of, AI could be premature and might be harmful at this time. We think it’s possible that by drawing attention to the nexus of security and emerging technologies (including AI), CSET could lead to premature regulatory attention and thus to harm. However, we believe CSET shares our interest in caution on this front and is well-positioned to communicate carefully.

CSET indeed communicated carefully and with nuance most of the time, at least according to my reading of its testimonials to the US Congress. In particular, it seemed likely that the late Trump administration was going to take punitive actions against China, and providing expert considerations on CSET’s area of expertise seemed unlikely to have done harm. There could be some scenarios in which any testimony at all increases political tensions, but this seems unlikely. However, some of the positions which CSET advocated for, e.g., openness and taking in top foreign talent from China, do map clearly across partisan lines, and if that proportion exceeds some threshold, or if CSET never gives support to uniquely Republican stances, CSET and the positions it defends might eventually come to be perceived as partisan.

With regards to research, CSET appears to be extremely prolific, per CSET’s list of publications. Some publications which appeared particularly relevant for evaluation purposes are:

Interestingly, CSET’s model of working within the prestigious mainstream seems to be particularly scalable, in a way which other organizations in the longtermist sphere are not. That is, because CSET doesn’t specifically look for EAs when hiring, CSET’s team has been able to quickly grow. This is in comparison with, for example, an organization like Rethink Priorities. The downside of this is that hires might not be aligned with longtermist interests.

Besides testimonials and research, CSET also has a large number of media appearances (cset.georgetown.edu/article/cset-experts-in-the-news through cset.georgetown.edu/article/cset-experts-in-the-news-10). I’m inclined to think that these appearances also have some kind of positive impact, though I am again uncertain of their magnitude.

CSET also carries out a large number of translations of Chinese policy and strategy documents. Lastly, I also occasionally encounter CSET’s research “in the wild”, e.g., these two blog posts by Bruce Schneier, a respected security expert, mentios a CSET report. This is at least some evidence that relevant experts read these.

Overall, the work that I have read appears to be lucid. But my knowledge of US policy work impact pathways is particularly fuzzy, and the pathways to influence policy are themselves fuzzy and uncertain. Further, unlike with some other organizations, there isn’t an annual review I can bootstrap an evaluation from.

For this reason, it is particularly tempting for me to defer to an outside view, like OpenPhilanthropy’s grant rationale for the creation of CSET, and its willingness to donate an initial $55 million in 2019, and an additional $8 million at the beginning of 2021. If OpenPhil hadn’t been willing to continue to fund CSET, I’d still guess that CSET’s work was valuable, but I would be fairly uncertain as to whether it was a comparatively good bet.

In conclusion, CSET’s work seems within what I would expect a competent think tank would produce. Given that OpenPhilanthropy is still funding them, I expect them to still be valuable. In particular, its think-tank model seems particularly scalable.

Sources


Future of Life Institute (FLI)

Epistemic status for this section: Uncertain about object-level facts regarding FLI.

Questions

Tentative answers

FLI was also briefly covered by Larks. I think Wikipedia does a better job summarizing FLI than the FLI website:

The Future of Life Institute (FLI) is a nonprofit research institute and outreach organization in the Boston area that works to mitigate existential risks facing humanity, particularly existential risk from advanced artificial intelligence (AI). Its founders include MIT cosmologist Max Tegmark and Skype co-founder Jaan Tallinn, and its board of advisors includes entrepreneur Elon Musk.

Some notable past activities include organizing conferences—such as the Asilomar Conference, which produced the Asilomar Principles on beneficial AI—work on Lethal Autonomous Weapons Systems, giving out the future of life award, and general policy work (open letters, initiatives, pledges, video content, podcasts, etc.) FLI is also a giving vehicle, and recently announced a $25M grant program financed by Vitalik Buterin. The Centre for the Governance of AI thanks FLI on its annual report.

To pick an example, for their work on Lethal Autonomous Weapons Systems, their model of impact seems to be that by raising awareness of the topic through various activities, and by pushing governments, NGOs and supranational organizations, they could institute a ban on Lethal Autonomous Weapons. This attempt would also act as a test-ground for “AI Arms Race Avoidance & Value Alignment.” So far, while they have raised awareness of the topic, a ban doesn’t seem to be forthcoming. Their video on slaughterbots reached a million views on youtube, but, per Seth Baum’s talk in EA Global 2018, “the video was fairly poorly received by a lot of important people in international security policy communities, and because of that it has made it more difficult for the people behind the video to get their message out there to these very important audiences.”

The core team mentioned in their webpage had just seven members, but increased to nine as I was writing this piece. Of these nine, five mention other current affiliations, and it’s unclear how many full-time equivalents FLI currently employs. In particular, I’d expect that to make inroads on their five core issues mentioned in their website (x-risk, artificial intelligence, nuclear weapons, biotechnology and climate change), a larger team would be needed.

In short, I’m uncertain about how valuable policy work is, about how valuable the specific policy work which FLI has done is, and about whether FLI intends to continue doing policy work. Colleagues have mentioned that FLI isn’t so much an organization as “a hat which sometimes people wear,” which seems plausible.


LessWrong

Epistemic status: The graphs serve as a sanity check on my intuitions, rather than being their main drivers.

Questions

Tentative answers

As I understand it, LessWrong’s benefits are

See here and here for other people using different wording.

With regards to concrete research outputs produced or catalyzed, some recent examples in the last three months from the list of curated posts are related to AI alignment are:

With regards to community building, some interaction happens in the comments. Further, the LessWrong team organizes activities, like Solstice celebrations, Petrov Day games, talks, etc. One rough measure of the community building aspect could be the number of new users with more than 500 or 1000 karma in the last couple of years. If we search for these, we find the following:

Note that this is, in a sense, unfair to recent years, because newly active users haven’t had time to accumulate as much karma as old users. Nonetheless, the conclusion that the LW community recovered from its previous decline holds.

It’s unclear to me exactly how valuable the production of around 10 highly engaged users with the rationality community is, but the intellectual output of those new 10 users seems probably comparable to that of a small or medium-sized research institute. And the combined output of LW seems much greater. Also note that this would be 10 new highly active users per year.

To the extent that these new users belong to already established organizations and just share the output of their work on LessWrong, LessWrong also seems valuable as a locus of discussion. But this doesn’t seem to be the main driver of growth in highly engaged users; of the 14 users who joined since the beginning of 2019 and have accumulated more than 500 karma, only around 3 belong to EA-aligned organizations.

We can also analyze the number of posts above 100 votes per year, or the total number of votes given to posts in each year. I’m using number of votes (number of people who vote) instead of karma (which includes a multiplier) because the LW API makes that easier to get. In any case, we find

If, as a rough approximation, we take 100 votes (for posts) as equivalent to two researcher/weeks, 40,000 votes in 2020 would equal 200 researcher months, or 17 researcher/years.

A more qualitative approach would involve, e.g., looking at the LessWrong Review for 2018, and asking how much one would be willing to pay for the creation and curation of the collected posts, or comparing their value to the value of FHI’s publications for the same year. One would have to adjust for the fact that around ¼th of the most highly upvoted posts are written by MIRI employees.

In conclusion, LW seems to catalyze or facilitate a relatively large amount of research, and that it does so relatively efficiently, with around 6 FTEs (per the team page). Concretely, LessWrong appears to produce substantially more than one FTE worth of research per FTE. One key question is whether many of the LessWrong posts would have just been written elsewhere.

In addition, the LessWrong codebase is also used by the EA Forum and by the AI Alignment Forum.

Sources


Rethink Priorities (RP)

Epistemic status: Only talking about explicitly longermist-branded parts of their research.

Questions

Tentative answers

Rethink Priorities has recently been expanding into the longtermist sphere, and it did so by hiring Linch Zhang and Michael Aird, the latter part-time, as well as some volunteers/interns.

At this point, I feel that the number of longtermist FTEs is so small that I wouldn’t be evaluating an organization, I would be evaluating individuals. All in all, Zhang and Aird haven’t spent enough time at RP that I feel that their output would be representative. This is in contrast to, e.g., FHI’s Research Scholars program, which is large enough that I feel it would make more sense to talk about the average quality of a researcher. That said, some of RP’s recent inputs can be found under their EA Forum tag.

With regards to the expected quality of future research, on the one hand, past high quality research is predictive of future quality. On the other hand, research into invertebrate sentience feels foundational for animal-focused ethics and activism in a way which seems hard to upstage, so one might expect some regression to the mean.

Sources


Simon Institute for Long-Term Governance (SILG)

Epistemic status: Brief and cursory. Considerations apply to other new organizations.

Questions

Tentative answers

I imagine that the prior distribution of success for new organizations is pretty long-tailed (e.g., a Pareto distribution). This would lead to a high initial expected value for new organizations, which most of the time sharply drops off after some initial time has passed and there is more information about the promisingness of the project. I imagine that ~two years might be enough to determine if a new organization is promising enough to warrant further investment.

If that was the case, the optimal move would look like funding a lot of new organizations, most of which are then deprived of funding shortly after an initial grace period.

It’s not clear how to create a functional culture around that dynamic. Silicon Valley aguably seems to be able to make it work, but they have somewhat reliable proxies of impact (e.g., revenue, user growth), whereas long-termists would have to rely on uncertain proxies.

The above considerations are fairly generic, and would apply to organizations other than SILG.

Overall, I estimate that funding SILG for the first two years of existence and seeing how they fare seems valuable, but I’m not very certain.

Sources


80,000 hours

Epistemic status: Deferring a lot to 80,000h’s evaluation of itself.

Questions

Tentative answers

80,000 hours has a clear evaluation of itself. For me, the gist is that

  1. 80,000 hours appears to have reached a point of maturity: Each programme is working well on its own terms. There’s a sensible, intuitive case for why each should exist, and their mechanisms for impact seem reasonably solid. They all seem to generate a reasonable number of plan changes or other value, and I expect them to compare well with alternatives. Big picture, 80,000 Hours seems likely to be among the biggest sources of talent into longtermist EA over the last couple of years, and it seems great to capitalize on that.
  2. The CEO is keen on expanding:

“Two years ago, I felt more uncertain about cost effectiveness and was more inclined to think we should focus on improving the programmes. My views feel more stable now, in part because we’ve improved our impact evaluation in response to critical feedback from 2018, clarified our views on the one-on-one programmes, and taken steps to limit negative side effects of our work. So, I think it makes sense to shift our focus toward growing the programmes' impact. Below I propose a two-year growth plan in which we aim to add 4.5 FTE in 2021, and 7.5 in 2022, though we plan to fundraise for 3.75 and 6.5, as we expect to hire no more than that many over the next two years in practice.”

Now, normally I’d think that the key questions were something like:

And indeed, most of 80,000 hours' impact tracking and quantification is done with regards to career plan changes (operationalized as “discounted, impact-adjusted peak years”). However, per the 80,000 hours review:

We remain unsure that plan changes are the best framework for thinking about 80,000 Hours' impact, and we think they capture only a minority of the value, especially for the website and podcast. For example, I think it’s plausible that most of our past impact has come from getting the EA movement more focused on longtermism and spreading other important ideas in society. An analysis I did this year confirmed my previous impression that 80,000 Hours is among the biggest and most effective ways of telling people about EA (though I expect less cost effective than the most successful written content, such as Doing Good Better and Slate Star Codex).

It is possible that further estimation of non-career plan change related impact would be clarifying, even if the estimation is very fuzzy. In particular, to the extent that most of 80,000 hours' impact comes from influencing the EA community, and this sounds plausible, having most of their evaluation focus on career plan changes feels misguided (cf. Streetlight effect).

(Despite feeling comfortable with the guess above, in practice, I’ve found that estimating total impact by estimating the impact of a measurable part and the fraction of value it represents leads to large errors)

With regards to cost-efficiency, 80,000 hours had a budget in 2020 of approximately $3M, and around 19 FTEs.

In short, 80,000 hours' career changes seem valuable, but most of the organization’s impact might come from fuzzier pathways, such as moving the EA community and 80,000 hours' followers in a more longtermist direction. I’m uncertain about the value of expansion.

Sources


Observations

I don’t have any overarching conclusions, so here are some atomic observations:

I’d welcome comments about the overall method, about whether I’m asking the right questions for any particular organization, or about whether my tentative answers to those questions are correct, and about whether this kind of evaluation seems valuable. For instance, it’s possible that I would have done better by evaluating all organizations using the same rubric (e.g., leadership quality, ability to identify talent, working on important problems, operational capacity, etc.)

I’d also be interested in algorithms to allocate funding supposing one had answers to all the questions I pose above, but did not have a clear way of comparing the impact of organizations working on different domains.

Thanks to Ozzie Gooen, Gustavs Zilgavis, Kelsey Rodriguez, Tegan McCaslin for comments and suggestions.

Appendix: Organizations about whose evaluations I’m less sure

Center on Long-term Risk (CLR)

Epistemic status for this section: Confused. In particular, I get the sense that for CLR, more than for other organizations, a fair evaluation probably requires deeply understanding what they do, which I don’t.

Questions

Tentative answers

Previously, Larks briefly reviewed CLR on his 2020 AI Alignment Literature Review and Charity Comparison. Sadly, CLR’s work on AI Safety related problems seems hard to judge as an outsider on the merits, and I get the impression that they are fairly disconnected from other longtermist groups (though CLR moved to London last year, which might remedy this.) This Alignment Forum post makes the case that multi-agent reinforcement learning, which CLR plans to explore in 2021, isn’t particularly neglected. Their Cooperation, Conflict, and Transformative Artificial Intelligence: A Research Agenda series on the Alignment forum didn’t get many comments.

Fortunately, one of CLR’s aims for the year is to “elicit feedback from outside experts to assess the quality and impact of our work”; I’m curious to see how that goes.

I’m not sure about whether further work on malevolence would be fruitful. In particular, it seems to me that the original post was very interesting and engaging. However, possible conclusions or proposals stemming from this kind of project are probably not implementable in the current political system. For instance, requiring psychopathy tests for politicians, or psychological evaluation, seems very unrealistic.

That said, perhaps one possible longer-term strategy might be to have proposals ready which can be implemented in the ensuing policy window following unexpected turmoil (e.g., pushing for psychopathy tests for politicians might have been more feasible in the aftermath of the Nürnberg trials, or after Watergate.) I imagine that people who interface with policy directly probably have better models about the political feasibility of anti-malevolence proposals.

Maybe considering CLR’s research agenda isn’t a good way to think about its potential impact. Daniel Kokotajlo’s work on AI timelines strikes me as valuable, and is outside that research agenda.

I have the subjective impression that CLR has historically been good at providing mentorship/funding for junior people trying to jump into EA research, e.g., for Michael Aird, Jaime Sevilla, even when their ethics were not particularly suffering-focused.

I found CLR particularly transparent with respect to their budget; their expected budget for 2021 was $1,830,000, and they expect to have 13.7 FTEs for the year. Commenters pointed out that this was surprisingly large compared to other organizations, e.g., 80,000 hours has around 19 FTEs (on a ~$3M budget).

In short, I don’t feel particularly enthused about their research agenda, but overall I’m not sure how to think about CLR’s impact.

Sources


Future of Humanity Institute

Epistemic status for this section: Arguably shouldn’t exist; FHI was just too large to be evaluated in a short time, so instead I rely mostly on status as a lagging indicator of impact.

Questions

Tentative answers

Per their team page, FHI is divided into the following teams:

Despite living under the FHI umbrella, each of these projects has a different pathway to impact, and thus they should most likely be evaluated separately. Note also that, unlike most other groups, FHI doesn’t really have consistent impact accounting for the organization as a whole. For instance, their last quarterly report, from their news section is from January to March 2020 (though it is possible that they have yet to publish their annual review for 2020.)

Consider in comparison 80,000 hours' annual review, which outlines what the different parts of the organization are doing, and why each project is probably valuable. I think having or creating such an annual review probably adds some clarity of thought when choosing strategic decisions (though one could also cargo-cult such a review solely in order to be more persuasive to donors), and it would also make shallow evaluations easier.

In the absence of an annual review to build upon, I’m unsatisfied with my ability to do more than a very shallow review in a short amount of time. In particular, I start out with the strong prior that FHI people are committed longtermists doing thoughtful work, and browsing through their work doesn’t really update me much either against or in favor.

I imagine that this might change as I think more about this, and maybe come up with an elegant factorization of FHI’s impact. In any case, below are some notes on each of the groups which make up FHI.

In the meantime, it seems that FHI doesn’t seem to be hurting for money, but that Open Phil is hesitant to donate too much to any particular organization. If one thinks that appeasing Open Phil’s neurosis is particularly important, which, all things considered, might be, or if one thinks that FHI is in fact hurting for money, FHI might be a good donation target.

Macrostrategy and AI Safety Research Groups

Some of the outputs from these two groups were favorably reviewed by Larks here.

Biosecurity Research Group

Some publications can be found in FHI’s page for the research group’s members (Gregory Lewis, Cassidy Nelson, Piers Millett). Gregory Lewis also has some blog posts on the EA forum.

I browsed their publications, but I don’t think I’m particularly able to evaluate them, given that they are so far outside my area of expertise. In the medium term (e.g., once the pandemic has subsided), some outside expert evaluation in Open Philanthropy’s style might be beneficial.

Nonetheless, I’m somewhat surprised by the size of the team. In particular, I imagine that to meaningfully reduce bio-risk, one would need a bigger team. It’s therefore possible that failing to expand is a mistake. However, commenters on a draft of this post pointed out that this isn’t straightforward; expanding is difficult, and brings its own challenges.

Centre for the Governance of AI (GovAI)

Some of the outputs from the Centre for the Governance of AI were favorably reviewed by Larks here (same link as before).

In addition, GovAI has its own 2020 Annual Report. It also has a post on the EA forum outlining its theory of impact, which is outlined with extreme clarity.

Research Scholars Programme, DPhil Scholars

A review of FHI’s Research Scholars Programme can be found here. The page for the DPhil Scholarship can be found here. FHI also has a Summer Research Fellowship, a review of which can be found here.

Overall, I’d guess that these programs have similar pathways to impact to some of the LTF grants to individual researchers, but the advantage that the participants gain additional prestige through their association with Oxford (as in the case of Research Scholars), or become more aligned with longtermist priorities (perhaps as in the case of the DPhil program).

Other associates and affiliates.

Associates and affiliates could contribute a small but significant part of FHI’s impact, but in the absence of very detailed models, I’m inclined to consider them as a multiplier (e.g. between x 1.05 and x 1.5 on FHI’s base impact, whatever that may be).

Conclusion

In conclusion, FHI’s output is fairly large and difficult to evaluate, particularly because they don’t have a yearly review or a well organized set of outputs I can bootstrap from. GovAI seems to be doing particularly valuable work. I still think highly of the organization, but I notice that I’m relying on status as a lagging indicator of quality.

Sources


Global Priorities Institute

Epistemic status: Uncertain about how valuable GPI’s work is, and about my ability to evaluate them.

Questions

Tentative answers

GPI’s 2020 annual report is fairly short and worth reading in full.

It describes GPI’s aims as:

The Global Priorities Institute (GPI) exists to develop and promote rigorous academic research into issues that arise in response to the question “What should an agent do with a given amount of resources, insofar as her aim is to do the most good?”. The investigation of these issues constitutes the enterprise that we call global priorities research. It naturally draws upon central themes in (in particular) the fields of economics and philosophy; the Institute is interdisciplinary between these two academic fields.

Overall, I see various pathways to impact which could arise from this kind of philosophy work:

  1. Philosophical clarity might be needed to optimally allocate donations. At the donation volume of an organization like OpenPhilanthropy or the Gates Foundation, relatively subtle changes in philosophical stances could lead to large changes in funding allocation. Further, some empirical considerations, such as those relating to the hinge of history hypothesis could also have more than marginal impact.
  2. Academic consensus could lead to policy change, by building the philosophical backbone of longtermism which would support and allow for future policy work.
  3. In particular, acquiring prestige in an academic field to then later influence policy may not require the academic field to be useful (i.e., it could be prestige about abstruse philosophical disputes). For example, testimony on future generations to the UK Parliament by an Oxford professor may be listened to because of the Oxford professorship, independent of its field.
  4. Trailblazing philosophy might pave the way for future practical developments. Exploring the moral landscape could lead to understanding the shape of our values, and realizing that e.g., invertebrates may hold some moral weight, or that most of the value of humanity may lie in its far away future. Organizations could later be created to work on the issues identified. A particularly striking example of this might be Trammell’s work on patient philanthropy, which might lead to a Patient Philanthropy fund. Another example might be Brian Tomasik’s essays on reducing suffering.
  5. Good philosophy might facilitate movement building, particularly inside academia. For instance, university professors might give courses on longtermism.
  6. Understanding ethical truths and decision theories at an extreme level of clarity would allow for the development of safer AI. This doesn’t seem to be GPI’s focus.

It is possible that I am missing some paths to impact. Right now, I see GPI as mostly aiming for 2., and growing its contingent of economists to allow for 3. 5. also seems to be happening, but it’s unclear what role GPI plays there (though potentially it could be a substantial role).

Readers might want to browse GPI’s list of publications (note that the list also contains papers which are relevant to GPI’s research agenda by authors not affiliated with GPI). I’m personally confused about their object level value, though some people I respect tell me that some are great.

In short, I’m fairly uncertain about GPI’s pathway to impact. Acquiring prestige and status might enable future policy work. Economics research, which GPI has been expanding into, seems more valuable.

Sources


Notes

[1]. One common unit might be “Quality-Adjusted Research Projects'‘, which could capture how efficiently an organization produces valuable research. However, that unit might be unsatisfactory, because research in different areas probably leads to differentially different outcomes. A different unit might be a "microtopia”, which according to oral lore was defined by Owen Cotton-Barratt to represent one millionth of the value of an ideal long-termist utopia. One might also try to compare the value of additional funding to a threshold, like the value of OpenPhilanthropy’s last (longtermist) dollar, or to compare to a given level of formidability.

[2]. Initially, I thought that the result of this project might be a GiveWell-style evaluation of longtermist organizations, just many, many orders of magnitude more uncertain. For instance, if organization A produces between 1 and 106 “utilons'‘ per unit of resources (attention, effort, money, etc.), and organization B produces between 0.01 and 103 "utilons” per unit of resources, we would want to choose organization A over organization B, even though the impact estimates overlap and are very uncertain.

[3]. Below is a list of perhaps notable organizations which I could have evaluated but didn’t. As mentioned, because of their additional complexity, and to bound the scope of this post, I decided to exclude meta organizations.

[4]. Which suggests a bias, perhaps because I’m reticent to assign probabilities lower than 1%, even if it’s per year. In the estimates later in the section, I ended up going mostly with yearly estimates based on my 100 year estimates.

[5].Michael Air’d Database of existential risk estimates.

[6]. Manhattan Project. “The Manhattan Project began modestly in 1939, but grew to employ more than 130,000 people and cost nearly US$2 billion (equivalent to about $23 billion in 2019).”

[7]. Lockheed Martin F-35 Lightning II development. “The program received considerable criticism for cost overruns during development and for the total projected cost of the program over the lifetime of the jets. By 2017 the program was expected over its lifetime (until 2070) to cost $406.5 billion for acquisition of the jets and $1.1 trillion for operations and maintenance.”

[8]. general purpose grants are likely less valuable per dollar than the best way to spend the marginal dollar for longtermist impact.

[9]. For instance, Exceeding expectations: stochastic dominance as a general decision theory makes the point that stochastic dominance (A stochastically dominates B if 1) for all events x the probability for equal or better events is greater or equal in A than in B, and 2) there is at least one possible event for which the inequality is strict) generalizes even to comparisons of events with infinite or undefined expected value. Further, in the presence of “background uncertainty”, stochastic dominance provides similar results to expected value, which might convince expected value skeptics to take some Pascalian-seeming wagers if the probability on which they depend is small, but not too small.

Note that the paper doesn’t word things that way. It also suggests in the latter sections that stochastic dominance stands as a decision theory on its own, which I’m very skeptical about.