Skip to main content

Stephen Hawking: Superhumans Able to Manipulate Their Own DNA Would Replace Us

Penelitian - Stephen Hawking, the physicist whose bodily paralysis turned him into a symbol of the soaring power of the human mind, feared a race of “superhumans” capable of manipulating their own evolution.

Before he died in March, the Cambridge University professor predicted that people this century would gain the capacity to edit human traits such as intelligence and aggression. And he worried that the capacity for genetic engineering would be concentrated in the hands of the wealthy.

Penelitian Stephen Hawking Superhumans able to manipulate their own DNA would replace us

Hawking mulled this future in a set of essays and articles being published posthumously Tuesday as Brief Answers to the Big Questions, a postscript of sorts to his 1988 A Brief History of Time: From the Big Bang to Black Holes, which has sold more than 10 million copies.

An excerpt released two days in advance by the Sunday Times sheds light on the final musings of the physicist and bestselling author, who was beset by a degenerative motor neurone disease similar to amyotrophic lateral sclerosis, or Lou Gehrig’s disease. If the human race manages to redesign itself, it will probably spread out and colonise other planets and stars

“A new phase of what might be called self-designed evolution, in which we will be able to change and improve our DNA. We have now mapped DNA, which means we have read ‘the book of life’, so we can start writing in corrections,” Hawking wrote.

Initially, he predicted, these modifications would be reserved for the repair of certain defects, such as muscular dystrophy, that are controlled by single genes and thus make for relatively simple corrections. There would be an attempt to pass laws restricting the genetic engineering of human traits

“Nevertheless, I am sure that during this century people will discover how to modify both intelligence and instincts such as aggression. But some people will not be able to resist the temptation to improve human characteristics, such as size of memory, resistance to disease and length of life,” Hawking anticipated.

“Once such superhumans appear, there are going to be significant political problems with the unimproved humans, who will not be able to compete. Presumably, they will die out, or become unimportant,” Hawking reasoned.

“A race of self-designing beings who are improving themselves at an ever-increasing rate. If the human race manages to redesign itself, it will probably spread out and colonise other planets and stars,” Hawking envisioned.

In the excerpted material, Hawking did not elaborate on the disparities between “superhumans” and the “unimproved”, but an article accompanying the excerpt in the Sunday Times suggests that Hawking’s fear was specifically that “wealthy people will soon be able to choose to edit their own and their children’s DNA”.

The article, by the newspaper’s science editor, draws a parallel to the 20th-century eugenics movement – premised similarly on the notion that human improvement could arise from genetic manipulation.

Some researchers and ethicists already fear that DNA editing may be eclipsing existing standards. Apprehension mainly surrounds so-called CRISPR DNA segments, “clustered regularly interspaced sort palindromic repeats” which have evolved from a component of bacterial defence into a means of altering specified DNA sequences with an eye toward enhancement.

The gene-editing technology has already been a source of major patent disputes. There are hopes that the tool could be used for low-cost disease detection. Meanwhile, applications in the food industry have alarmed consumer and environmental groups.

Though Hawking saw himself as an “optimist,” as he wrote in Brief Answers to the Big Questions, the book is a warning of runaway technological ambition. He was particularly worried that the promise of artificial intelligence, the possibility of the eradication of disease and poverty, would blind its developers to long-term costs, namely that humans could lose control over AI’s growth.

“One can imagine such technology outsmarting financial markets, out-inventing human researchers, out-manipulating human leaders and potentially subduing us with weapons we cannot even understand,” Hawking wrote.

“Whereas the short-term impact of AI depends on who controls it, the long-term impact depends on whether it can be controlled at all. Artificial intelligence that is able to improve upon itself without human assistance may be able to amass intelligence that “exceeds ours by more than ours exceeds that of snails”, Hawking warned.

Hawking, whose exploration of gravity and black holes marked a transformation in modern physics, proffered answers to a number of other mysteries. Among them are the long-term feasibility of life on Earth (“I regard it as almost inevitable that either a nuclear confrontation or environmental catastrophe will cripple the Earth at some point in the next 1,000 years”); the existence of God (“If you like, you can call the laws of science ‘God’, but it wouldn’t be a personal God that you would meet and put questions to”); and the biggest threat to the future of the planet (“asteroid collision”).

Comments

Popular

Graphene Mechanical Pixels for Interferometric Modulator Displays

US Authorities Paves Way to Get Lab Meat on Plates

The Genetic Prehistory of the Andean Highlands 7000 Years BP Though European Contact

Spontaneous Innovation of Hook-bending and Unbending in Orangutans (Pongo abelii)

Cats Use Hollow Papillae to Wick Saliva Into Fur