Open Call for Advisees and Collaborators, May 2022

UPDATE: The open call for advisees and collaborators is now closed. Thank you to everyone who applied. However, anyone interested in seeking our advice and/or collaborating with us is still welcome to contact us as per the instructions below and we will include them in our next advisees and collaborators program.

GCRI is currently welcoming inquiries from people who are interested in seeking our advice and/or collaborating with us as part of our fourth annual Advising and Collaboration Program. Inquiries may cover any aspect of global catastrophic …

Read More »

Book Review: The Precipice

View the paper “Book review: The Precipice”

Book review of The Precipice: Existential Risk and the Future of Humanity, by Toby Ord, Hachette Books, 2020.

The new book The Precipice by Toby Ord provides a wide-ranging survey of topics related to global catastrophic risk. Compared to other books on global catastrophic risk, The Precipice stands out for its depth of discussion, its quality of scholarship, and its readability. However, the book errs in its emphasis on only the most extreme global catastrophe scenarios, its strong belief in the resilience of civilization, and …

Read More »

Pandemic Refuges: Lessons from Two Years of COVID-19

View the paper “Pandemic Refuges: Lessons from Two Years of COVID-19”

Refuges have been proposed as a means of ensuring that at least some people survive a global catastrophe. While it would be better to avoid the catastrophe in the first place, if a catastrophe is to occur, a refuge could be a real difference-maker in terms of the long-term effects on human civilization. Prior refuges research emphasizes highly isolated locations such as underground, underwater, or in outer space. These exotic concepts may seem far removed …

Read More »

March Newsletter: Implications of the War in Ukraine

Dear friends,The Russian invasion of Ukraine is already proving to be an event of profound importance for global catastrophic risk. As detailed in the GCRI Statement on the Russian Invasion of Ukraine, the war’s implications for nuclear war risk are especially strong, but it also has implications for other risks including climate change, pandemics, and artificial intelligence. These changes are coming from the war itself and from the accompanying shifts in global politics. We at GCRI hope that the war can reach a prompt and peaceful …

Read More »

Early Reflections and Resources on the Russian Invasion of Ukraine

View the article “Early Reflections and Resources on the Russian Invasion of Ukraine”.

This article, published in the Effective Altruism Forum, presents analysis of the Russian invasion of Ukraine written for a global catastrophic risk audience. The article discusses nuclear war risk, the changing geopolitical landscape, and recommendations for personal preparedness and philanthropy. It also describes the author’s own activities in addressing the immediate risk and presents a compilation of resources for learning more about the war.

See also the GCRI Statement on the Russian Invasion of Ukraine.

The …

Read More »

How to Evaluate the Risk of Nuclear War

View the article “How to evaluate the risk of nuclear war”.

This article, published in BBC Future, discusses the quantitative analysis of nuclear war risk. It is written in the context of the Russian invasion of Ukraine and also discusses more general analytical issues, such as found in GCRI’s nuclear war research.

See also the GCRI Statement on the Russian Invasion of Ukraine.

The article begins as follows:

One day last week, I woke up in the morning and looked out the window to see the Sun was shining. My neighbourhood …

Read More »

Recommendations to the OSTP on the National Artificial Intelligence Research and Development Strategic Plan

View GCRI’s submission to the OSTP on the National Artificial Intelligence Research and Development Strategic Plan.

On 1 January 2021, the National AI Initiative Act of 2020 became law as part of the National Defense Authorization Act. The National AI Initiative Act calls for regular updates to the National AI R&D Strategic Plan to include “goals, priorities, and metrics for guiding and evaluating how the agencies carrying out the National AI Initiative will support AI research. The Office of Science and Technology Policy (OSTP) requested input from the …

Read More »

GCRI Statement on the Russian Invasion of Ukraine

The ongoing Russian invasion of Ukraine is already shaping up to be an event of historic proportions. This includes, but is certainly not limited to, its implications for global catastrophic risk. We at GCRI are monitoring the unfolding events with great concern. While it is always important to understand all parties’ perspectives on a conflict, in this case we find ourselves strongly condemning the actions of the Russian government. Our hearts go out to the many people in Ukraine who have been tragically affected by …

Read More »

February Newsletter: Ukraine & Pluralism

Dear friends,

We at GCRI are watching the ongoing Russian invasion of Ukraine with great concern. In addition to the grave harm being inflicted on the Ukrainian people, this invasion also constitutes a large escalation of tensions between Russia and the West and a shooting war adjacent to several NATO countries. In our judgment, this increases the risk of US-Russia or NATO-Russia nuclear war and accompanying nuclear winter. Our hearts go out to the people of Ukraine who are enduring this tragic violence. For the sake …

Read More »

GCRI Statement on Pluralism in the Field of Global Catastrophic Risk

Global catastrophic risk is an important societal issue area. As such, it is to be expected that there will be a variety of views on it. We at GCRI believe that it is important to consider a range of views to better understand the topic of global catastrophic risk and the constructive options for addressing the risk. We are likewise interested in supporting a pluralistic field of global catastrophic risk.

Many types of pluralism can be valuable for the field of global catastrophic risk. Examples include …

Read More »