End Of Humanity
   HOME

TheInfoList



OR:

Human extinction, also known as omnicide, is the hypothetical end of the human species due to either natural causes such as
population decline A population decline (also sometimes called underpopulation, depopulation, or population collapse) in humans is a reduction in a human population size. Over the long term, stretching from prehistory to the present, Earth's total human population ...
from sub-replacement fertility, an asteroid impact, or large-scale volcanism, or to
anthropogenic Anthropogenic ("human" + "generating") is an adjective that may refer to: * Anthropogeny, the study of the origins of humanity Counterintuitively, anthropogenic may also refer to things that have been generated by humans, as follows: * Human im ...
(human) causes. For the latter, some of the many possible contributors include climate change, global nuclear annihilation, biological warfare, and ecological collapse. Other scenarios center on emerging technologies, such as
advanced artificial intelligence Strong artificial intelligence may refer to: "Strong Artificial Intelligence (AI) is an artificial intelligence that constructs mental abilities, thought processes, and functions that are impersonated from the human brain. It is more of a phil ...
, biotechnology, or self-replicating nanobots. The scientific consensus is that there is a relatively low risk of near-term human extinction due to natural causes. The likelihood of human extinction through humankind's own activities, however, is a current area of research and debate.


History of thought


Early history of thinking about human extinction

Before the 18th and 19th centuries, the possibility that humans or other organisms could become extinct was viewed with scepticism. It contradicted the principle of plenitude, a doctrine that all possible things exist. The principle traces back to Aristotle, and was an important tenet of Christian theology. Ancient Western philosophers such as Plato, Aristotle, and Lucretius wrote of the end of humankind only as part of a cycle of renewal. Later philosophers such as Al-Ghazali, William of Ockham, and Gerolamo Cardano expanded the study of logic and probability and began discussing abstract ''possible'' worlds, including a world without humans. Physicist Edmond Halley stated that the extinction of the human race may be beneficial to the future of the world. The notion that species can become extinct gained scientific acceptance during the Age of Enlightenment in the 17th and 18th centuries, and by 1800
Georges Cuvier Jean Léopold Nicolas Frédéric, Baron Cuvier (; 23 August 1769 – 13 May 1832), known as Georges Cuvier, was a French natural history, naturalist and zoology, zoologist, sometimes referred to as the "founding father of paleontology". Cuvier ...
had identified 23 extinct prehistoric species. The doctrine was further gradually undermined by evidence from the natural sciences, particular the discovery of fossil evidence of species that appeared to no longer exist, and the development of theories of evolution. In '' On the Origin of Species'',
Darwin Darwin may refer to: Common meanings * Charles Darwin (1809–1882), English naturalist and writer, best known as the originator of the theory of biological evolution by natural selection * Darwin, Northern Territory, a territorial capital city i ...
discussed the extinction of species as a natural process and core component of natural selection. Notably, Darwin was skeptical of the possibility of sudden extinctions, viewing it as a gradual process. He held that the abrupt disappearances of species from the fossil record were not evidence of catastrophic extinctions, but rather were a function of unrecognised gaps in the record. As the possibility of extinction became more widely established in the sciences, so did the prospect of human extinction. In the 19th century, human extinction became a popular topic in science (e.g., Thomas Robert Malthus's ''
An Essay on the Principle of Population An, AN, aN, or an may refer to: Businesses and organizations * Airlinair (IATA airline code AN) * Alleanza Nazionale, a former political party in Italy * AnimeNEXT, an annual anime convention located in New Jersey * Anime North, a Canadian an ...
'') and fiction (e.g., Mary Shelley's '' The Last Man''). In 1863, a few years after Charles Darwin published '' On the Origin of Species'',
William King William King may refer to: Arts *Willie King (1943–2009), American blues guitarist and singer *William King (author) (born 1959), British science fiction author and game designer, also known as Bill King *William King (artist) (1925–2015), Ame ...
proposed that Neanderthals were an extinct species of the genus '' Homo''. The
Romantic Romantic may refer to: Genres and eras * The Romantic era, an artistic, literary, musical and intellectual movement of the 18th and 19th centuries ** Romantic music, of that era ** Romantic poetry, of that era ** Romanticism in science, of that e ...
authors and poets were particularly interested in the topic. Lord Byron wrote about the extinction of life on earth in his 1816 poem " Darkness", and in 1824 envisaged humanity being threatened by a comet impact, and employing a missile system to defend against it. Mary Shelley's 1826 novel '' The Last Man'' is set in a world where humanity has been nearly destroyed by a mysterious plague. At the turn of the 20th century, Russian cosmism, a precursor to modern transhumanism, advocated avoiding humanity's extinction by colonizing space.


Atomic era

The invention of the atomic bomb prompted a wave of discussion about the risk of human extinction among scientists, intellectuals, and the public at large. In a 1945 essay, Bertrand Russell wrote that " e prospect for the human race is sombre beyond all precedent. Mankind are faced with a clear-cut alternative: either we shall all perish, or we shall have to acquire some slight degree of common sense." In 1950, Leo Szilard suggested it was technologically feasible to build a cobalt bomb that could render the planet unlivable. A 1950 Gallup poll found that 19% of Americans believed that another world war would mean "an end to mankind". Rachel Carson's 1962 '' Silent Spring'' raised awareness of environmental catastrophe. In 1983, Brandon Carter proposed the Doomsday argument, which used Bayesian probability to predict the total number of humans that will ever exist. The discovery of " nuclear winter" in the early 1980s, a specific mechanism by which nuclear war could result in human extinction, again raised the issue to prominence. Writing about these findings in 1983,
Carl Sagan Carl Edward Sagan (; ; November 9, 1934December 20, 1996) was an American astronomer, planetary scientist, cosmologist, astrophysicist, astrobiologist, author, and science communicator. His best known scientific contribution is research on ext ...
argued that measuring the severity of extinction solely in terms of those who die "conceals its full impact", and that nuclear war "imperils all of our descendants, for as long as there will be humans."


Post Cold War

John Leslie's 1996 book ''The End of The World'' was an academic treatment of the science and ethics of human extinction. In it, Leslie considered a range of threats to humanity and what they have in common. In 2003, British Astronomer Royal Sir Martin Rees published '' Our Final Hour'', in which he argues that advances in certain technologies create new threats for the survival of humankind and that the 21st century may be a critical moment in history when humanity's fate is decided. Edited by
Nick Bostrom Nick Bostrom ( ; sv, Niklas Boström ; born 10 March 1973) is a Swedish-born philosopher at the University of Oxford known for his work on existential risk, the anthropic principle, human enhancement ethics, superintelligence risks, and the rev ...
and
Milan M. Ćirković Milan M. Ćirković (born 11 March 1971 in Belgrade) is a Serbian astronomer, astrophysicist, philosopher and science book author. He has worked in the fields of astrobiology, global catastrophic risks and future of humanity where he also co-author ...
, ''
Global Catastrophic Risks A global catastrophic risk or a doomsday scenario is a hypothetical future event that could damage human well-being on a global scale, even endangering or destroying modern civilization. An event that could cause human extinction or permanen ...
'' was published in 2008, a collection of essays from 26 academics on various global catastrophic and existential risks. Toby Ord's 2020 book '' The Precipice: Existential Risk and the Future of Humanity'' argues that preventing existential risks is one of the most important moral issues of our time. The book discusses, quantifies, and compares different existential risks, concluding that the greatest risks are presented by unaligned artificial intelligence and biotechnology.


Causes

Potential anthropogenic causes of human extinction include global thermonuclear war, deployment of a highly effective biological weapon, an ecological collapse, runaway artificial intelligence, runaway
nanotechnology Nanotechnology, also shortened to nanotech, is the use of matter on an atomic, molecular, and supramolecular scale for industrial purposes. The earliest, widespread description of nanotechnology referred to the particular technological goal o ...
(such as a grey goo scenario), a scientific accident involving a micro black hole or vacuum metastability disaster, overpopulation and increased consumption pose the risk of resource depletion and a concomitant population crash,
population decline A population decline (also sometimes called underpopulation, depopulation, or population collapse) in humans is a reduction in a human population size. Over the long term, stretching from prehistory to the present, Earth's total human population ...
by choosing to have fewer children, displacement of naturally evolved humans by a new species produced by
genetic engineering Genetic engineering, also called genetic modification or genetic manipulation, is the modification and manipulation of an organism's genes using technology. It is a set of technologies used to change the genetic makeup of cells, including t ...
or technological augmentation. Natural and external extinction risks include high-fatality-rate
pandemic A pandemic () is an epidemic of an infectious disease that has spread across a large region, for instance multiple continents or worldwide, affecting a substantial number of individuals. A widespread endemic (epidemiology), endemic disease wi ...
, supervolcanic eruption, asteroid impact, nearby
supernova A supernova is a powerful and luminous explosion of a star. It has the plural form supernovae or supernovas, and is abbreviated SN or SNe. This transient astronomical event occurs during the last evolutionary stages of a massive star or when ...
or gamma-ray bursts, extreme solar flare, or alien invasion. Without intervention by unexpected forces, the
stellar evolution Stellar evolution is the process by which a star changes over the course of time. Depending on the mass of the star, its lifetime can range from a few million years for the most massive to trillions of years for the least massive, which is cons ...
of the Sun is expected to make Earth uninhabitable, then destroy it. Depending on its ultimate fate, the entire universe may eventually become uninhabitable.


Probability


Natural vs. anthropogenic

Experts generally agree that anthropogenic existential risks are (much) more likely than natural risks. A key difference between these risk types is that empirical evidence can place an upper bound on the level of natural risk. Humanity has existed for at least 200,000 years, over which it has been subject to a roughly constant level of natural risk. If the natural risk were high, then it would be highly unlikely that humanity would have survived as long as it has. Based on a formalization of this argument, researchers have concluded that we can be confident that natural risk is lower than 1 in 14,000 (and likely "less than one in 87,000") per year. Another empirical method to study the likelihood of certain natural risks is to investigate the geological record. For example, a comet or asteroid impact event sufficient in scale to cause an impact winter that would cause human extinction before the year 2100 has been estimated at one-in-a-million. Moreover, large
supervolcano A supervolcano is a volcano that has had an eruption with a Volcanic Explosivity Index (VEI) of 8, the largest recorded value on the index. This means the volume of deposits for such an eruption is greater than 1,000 cubic kilometers (240 cubic ...
eruptions may cause a volcanic winter that could endanger the survival of humanity. The geological record suggests that supervolcanic eruptions are estimated to occur on average about once every 50,000 years, though most such eruptions would not reach the scale required to cause human extinction. Famously, the supervolcano Mt. Toba may have almost wiped out humanity at the time of its last eruption (though this is contentious). Since anthropogenic risk is a relatively recent phenomenon, humanity's track record of survival cannot provide similar assurances. Humanity has only survived 77 years since the creation of nuclear weapons, and for future technologies, there is no track record at all. This has led thinkers like
Carl Sagan Carl Edward Sagan (; ; November 9, 1934December 20, 1996) was an American astronomer, planetary scientist, cosmologist, astrophysicist, astrobiologist, author, and science communicator. His best known scientific contribution is research on ext ...
to conclude that humanity is currently in a "time of perils" – a uniquely dangerous period in human history, where it is subject to unprecedented levels of risk, beginning from when humans first started posing risk to themselves through their actions.


Risk estimates

Given the limitations of ordinary observation and modeling, expert elicitation is frequently used instead to obtain probability estimates. In 2008, an informal survey of experts at a conference hosted by the Future of Humanity Institute estimated a 19% risk of human extinction by the year 2100, though, given the survey's limitations, these results should be taken "with a grain of salt".
Table source: Future of Humanity Institute, 2008.
There have been a number of other estimates of existential risk, extinction risk, or a global collapse of civilization: * Humanity has a 95% probability of being extinct in 7,800,000 years, according to J. Richard Gott's formulation of the controversial Doomsday argument, which argues that we have probably already lived through half the duration of human history. * In 1996, John Leslie estimated a 30% risk over the next five centuries (equivalent to around 9% per century, on average). * In 2003, Martin Rees estimated a 50% chance of collapse of civilisation in the twenty-first century. * The Global Challenges Foundation's 2016 annual report estimates an annual probability of human extinction of at least 0.05% per year. * A 2016 survey of AI experts found a median estimate of 5% that human-level AI would cause an outcome that was "extremely bad (e.g. human extinction)". * In 2020, Toby Ord estimates existential risk in the next century at "1 in 6" in his book '' The Precipice: Existential Risk and the Future of Humanity''. * Metaculus users currently estimate a 3% probability of humanity going extinct before 2100. * In a 2010 interview with '' The Australian'', Australian scientist Frank Fenner predicted the extinction of the human race within a century, primarily as the result of human overpopulation,
environmental degradation Environmental degradation is the deterioration of the environment (biophysical), environment through depletion of resources such as quality of air, water and soil; the destruction of ecosystems; habitat destruction; the extinction of wildlife; an ...
and climate change. * According to a 2020 study published in ''
Scientific Reports ''Scientific Reports'' is a peer-reviewed open-access scientific mega journal published by Nature Portfolio, covering all areas of the natural sciences. The journal was established in 2011. The journal states that their aim is to assess solely th ...
'', if deforestation and resource consumption continue at current rates, they could culminate in a "catastrophic collapse in human population" and possibly "an irreversible collapse of our civilization" in the next 20 to 40 years. According to the most optimistic scenario provided by the study, the chances that human civilization survives are smaller than 10%. To avoid this collapse, the study says, humanity should pass from a civilization dominated by the economy to a "cultural society" that "privileges the interest of the ecosystem above the individual interest of its components, but eventually in accordance with the overall communal interest." * Nick Bostrom, a philosopher at the University of Oxford known for his work on existential risk, argues that it would be "misguided" to assume that the probability of near-term extinction is less than 25% and that it will be "a tall order" for the human race to "get our precautions sufficiently right the first time", given that an existential risk provides no opportunity to learn from failure. * Philosopher John Leslie assigns a 70% chance of humanity surviving the next five centuries, based partly on the controversial philosophical doomsday argument that Leslie champions. Leslie's argument is somewhat frequentist, based on the observation that human extinction has never been observed, but requires subjective anthropic arguments. Leslie also discusses the anthropic
survivorship bias Survivorship bias or survival bias is the logical error of concentrating on entities that passed a selection process while overlooking those that did not. This can lead to incorrect conclusions because of incomplete data. Survivorship bias is ...
(which he calls an "observational selection" effect on page 139) and states that the '' a priori'' certainty of observing an "undisastrous past" could make it difficult to argue that we must be safe because nothing terrible has yet occurred. He quotes Holger Bech Nielsen's formulation: "We do not even know if there should exist some extremely dangerous decay of say the proton which caused the eradication of the earth, because if it happens we would no longer be there to observe it and if it does not happen there is nothing to observe." * Jean-Marc Salotti calculated the probability of human extinction caused by a giant asteroid impact. It is between 0.03 and 0.3 for the next billion years, if there is no colonization of other planets. According to that study, the most frightening object is a giant long period comet with a warning time of a few years only and therefore no time for any intervention in space or settlement on the Moon or Mars. The probability of a giant comet impact in the next hundred years is 2.2E-12.


Individual vs. species risks

Although existential risks are less manageable by individuals than – for example – health risks, according to Ken Olum, Joshua Knobe, and Alexander Vilenkin, the possibility of human extinction ''does'' have practical implications. For instance, if the "universal" Doomsday argument is accepted it changes the most likely source of disasters, and hence the most efficient means of preventing them. They write: "... you should be more concerned that a large number of asteroids have not yet been detected than about the particular orbit of each one. You should not worry especially about the chance that some specific nearby star will become a supernova, but more about the chance that supernovas are more deadly to nearby life than we believe."


Difficulty

Some scholars argue that certain scenarios such as global thermonuclear war would have difficulty eradicating every last settlement on Earth. Physicist Willard Wells points out that any credible extinction scenario would have to reach into a diverse set of areas, including the underground subways of major cities, the mountains of Tibet, the remotest islands of the South Pacific, and even to
McMurdo Station McMurdo Station is a United States Antarctic research station on the south tip of Ross Island, which is in the New Zealand-claimed Ross Dependency on the shore of McMurdo Sound in Antarctica. It is operated by the United States through the Unit ...
in Antarctica, which has contingency plans and supplies for long isolation. In addition, elaborate bunkers exist for government leaders to occupy during a nuclear war. The existence of nuclear submarines, which can stay hundreds of meters deep in the ocean for potentially years at a time, should also be considered. Any number of events could lead to a massive loss of human life, but if the last few (see minimum viable population) most resilient humans are unlikely to also die off, then that particular human extinction scenario may not seem credible.


Ethics


Value of human life

"Existential risks" are risks that threaten the entire future of humanity, whether by causing human extinction or by otherwise permanently crippling human progress. Multiple scholars have argued based on the size of the "cosmic endowment" that because of the inconceivably large number of potential future lives that are at stake, even small reductions of existential risk have great value. In one of the earliest discussions of ethics of human extinction,
Derek Parfit Derek Antony Parfit (; 11 December 1942 – 1 or 2 January 2017) was a British philosopher who specialised in personal identity, rationality, and ethics. He is widely considered one of the most important and influential moral philosophers of ...
offers the following thought experiment: The scale of what is lost in an existential catastrophe is determined by humanity's long-term potential – what humanity could expect to achieve if it survived. From a utilitarian perspective, the value of protecting humanity is the product of its duration (how long humanity survives), its size (how many humans there are over time), and its quality (on average, how good is life for future people). On average, species survive for around a million years before going extinct. Parfit points out that the Earth will remain habitable for around a billion years. And these might be lower bounds on our potential: if humanity is able to expand beyond Earth, it could greatly increase the human population and survive for trillions of years. The size of the foregone potential that would be lost, were humanity to become extinct, is very large. Therefore, reducing existential risk by even a small amount would have a very significant moral value.
Carl Sagan Carl Edward Sagan (; ; November 9, 1934December 20, 1996) was an American astronomer, planetary scientist, cosmologist, astrophysicist, astrobiologist, author, and science communicator. His best known scientific contribution is research on ext ...
wrote in 1983: "If we are required to calibrate extinction in numerical terms, I would be sure to include the number of people in future generations who would not be born.... (By one calculation), the stakes are one million times greater for extinction than for the more modest nuclear wars that kill "only" hundreds of millions of people. There are many other possible measures of the potential loss – including culture and science, the evolutionary history of the planet, and the significance of the lives of all of our ancestors who contributed to the future of their descendants. Extinction is the undoing of the human enterprise." Philosopher Robert Adams in 1989 rejects Parfit's "impersonal" views but speaks instead of a moral imperative for loyalty and commitment to "the future of humanity as a vast project... The aspiration for a better society – more just, more rewarding, and more peaceful... our interest in the lives of our children and grandchildren, and the hopes that they will be able, in turn, to have the lives of their children and grandchildren as projects." Philosopher
Nick Bostrom Nick Bostrom ( ; sv, Niklas Boström ; born 10 March 1973) is a Swedish-born philosopher at the University of Oxford known for his work on existential risk, the anthropic principle, human enhancement ethics, superintelligence risks, and the rev ...
argues in 2013 that preference-satisfactionist, democratic, custodial, and intuitionist arguments all converge on the common-sense view that preventing existential risk is a high moral priority, even if the exact "degree of badness" of human extinction varies between these philosophies. Parfit argues that the size of the "cosmic endowment" can be calculated from the following argument: If Earth remains habitable for a billion more years and can sustainably support a population of more than a billion humans, then there is a potential for 10 (or 10,000,000,000,000,000) human lives of normal duration.Parfit, D. (1984) Reasons and Persons. Oxford: Clarendon Press. pp. 453–454 Bostrom goes further, stating that if the universe is empty, then the accessible universe can support at least 10 biological human life-years; and, if some humans were uploaded onto computers, could even support the equivalent of 10 cybernetic human life-years. Some economists and philosophers have defended views, including exponential discounting and person-affecting views of population ethics, on which future people do not matter (or matter much less), morally speaking. While these views are controversial, even they would agree that an existential catastrophe would be among the worst things imaginable. It would cut short the lives of eight billion presently existing people, destroying all of what makes their lives valuable, and most likely subjecting many of them to profound suffering. So even setting aside the value of future generations, there may be strong reasons to reduce existential risk, grounded in concern for presently existing people. Beyond utilitarianism, other moral perspectives lend support to the importance of reducing existential risk. An existential catastrophe would destroy more than just humanity – it would destroy all cultural artifacts, languages, and traditions, and many of the things we value. So moral viewpoints on which we have duties to protect and cherish things of value would see this as a huge loss that should be avoided. One can also consider reasons grounded in duties to past generations. For instance,
Edmund Burke Edmund Burke (; 12 January NS.html"_;"title="New_Style.html"_;"title="/nowiki>New_Style">NS">New_Style.html"_;"title="/nowiki>New_Style">NS/nowiki>_1729_–_9_July_1797)_was_an_ NS.html"_;"title="New_Style.html"_;"title="/nowiki>New_Style"> ...
writes of a "partnership ... between those who are living, those who are dead, and those who are to be born". If one takes seriously the debt humanity owes to past generations, Ord argues the best way of repaying it might be to 'pay it forward', and ensure that humanity's inheritance is passed down to future generations. There are several economists who have discussed the importance of global catastrophic risks. For example,
Martin Weitzman Martin Lawrence Weitzman (April 1, 1942 – August 27, 2019) was an economist and a professor of economics at Harvard University. He was among the most influential economists in the world according to Research Papers in Economics (RePEc). His l ...
argues that most of the expected economic damage from climate change may come from the small chance that warming greatly exceeds the mid-range expectations, resulting in catastrophic damage. Richard Posner has argued that humanity is doing far too little, in general, about small, hard-to-estimate risks of large-scale catastrophes.


Voluntary extinction

Some philosophers adopt the
antinatalist Antinatalism or anti-natalism is the view that procreation is wrong. Antinatalists argue that humans should abstain from procreation because it is morally wrong. In scholarly and literary writings, various ethical arguments have been put forth i ...
position that human extinction would not be a bad thing, but a good thing. David Benatar argues that coming into existence is always serious harm, and therefore it is better that people do not come into existence in the future. Further, David Benatar, animal rights activist
Steven Best Steven Best (born December 1955) is an American philosopher, writer, speaker and activist. His concerns include animal rights, species extinction, human overpopulation, ecological crisis, biotechnology, liberation politics, terrorism, mass media ...
, and anarchist Todd May, posit that human extinction would be a positive thing for the other organisms on the planet, and the planet itself, citing, for example, the omnicidal nature of human civilization. The environmental view in favor of human extinction is shared by the members of Voluntary Human Extinction Movement who call for refraining from reproduction and allowing the human species to go peacefully extinct, thus stopping further
environmental degradation Environmental degradation is the deterioration of the environment (biophysical), environment through depletion of resources such as quality of air, water and soil; the destruction of ecosystems; habitat destruction; the extinction of wildlife; an ...
.


In fiction

Jean-Baptiste Cousin de Grainville's 1805 ''Le dernier homme'' (''The Last Man''), which depicts human extinction due to infertility, is considered the first modern apocalyptic novel and credited with launching the genre. Other notable early works include Mary Shelley's 1826 '' The Last Man'', depicting human extinction caused by a
pandemic A pandemic () is an epidemic of an infectious disease that has spread across a large region, for instance multiple continents or worldwide, affecting a substantial number of individuals. A widespread endemic (epidemiology), endemic disease wi ...
, and Olaf Stapledon's 1937 ''
Star Maker ''Star Maker'' is a science fiction novel by British writer Olaf Stapledon, published in 1937. The book describes a history of life in the universe, dwarfing in scale Stapledon's previous book, ''Last and First Men'' (1930), a history of the hu ...
'', "a comparative study of omnicide". Some 21st century pop-science works, including ''
The World Without Us ''The World Without Us'' is a 2007 non-fiction book about what would happen to the natural and built environment if humans suddenly disappeared, written by American journalist Alan Weisman and published by St. Martin's Thomas Dunne Books. It i ...
'' by Alan Weisman, and the TV specials '' Life After People'' and '' Aftermath: Population Zero'' pose a thought experiment: what would happen to the rest of the planet if humans suddenly disappeared? A threat of human extinction, such as through a technological singularity (also called an intelligence explosion), drives the plot of innumerable science fiction stories; an influential early example is the 1951 film adaption of '' When Worlds Collide''. Usually the extinction threat is narrowly avoided, but some exceptions exist, such as '' R.U.R.'' and
Steven Spielberg Steven Allan Spielberg (; born December 18, 1946) is an American director, writer, and producer. A major figure of the New Hollywood era and pioneer of the modern blockbuster, he is the most commercially successful director of all time. Spie ...
's ''
A.I. Artificial intelligence (AI) is intelligence—perceiving, synthesizing, and inferring information—demonstrated by machines, as opposed to intelligence displayed by animals and humans. Example tasks in which this is done include speech rec ...
''


See also


References


Sources

* * ** ** ** ** ** ** ** ** ** ** ** *
PDF
] * * *


Further reading

* * Jim Holt (philosopher), Holt, Jim, "The Power of Catastrophic Thinking" (review of Toby Ord, ''The Precipice: Existential Risk and the Future of Humanity'', Hachette, 2020, 468 pp.), '' The New York Review of Books'', vol. LXVIII, no. 3 (February 25, 2021), pp. 26–29. Jim Holt writes (p. 28): "Whether you are searching for a cure for cancer, or pursuing a scholarly or artistic career, or engaged in establishing more just institutions, a threat to the future of humanity is also a threat to the significance of what you do." * * * Ord, Toby (2020). '' The Precipice: Existential Risk and the Future of Humanity.'' Bloomsbury Publishing. * Torres, Phil. (2017). ''Morality, Foresight, and Human Flourishing: An Introduction to Existential Risks''. Pitchstone Publishing. . * Michel Weber, "Book Review: ''Walking Away from Empire''", ''Cosmos and History: The Journal of Natural and Social Philosophy'', vol. 10, no. 2, 2014, pp. 329–336.
What would happen to Earth if humans went extinct?
Live Science Live Science is a science news website run by Future via Purch, which it purchased from Imaginova in 2009. Stories and editorial commentary are typically syndicated to major news outlets, such as Yahoo!, MSNBC, AOL, and Fox News.{{fact, date=Marc ...
, August 16, 2020. {{Extinction