Yes, we are worried about the existential risk of artificial intelligence

Thus, in our view, Etzioni’s article distracts the reader from the core argument of the book and directs an ad hominem attack against Bostrom under the pretext of disputing his survey results. Gas exchange in the lungs We feel it is necessary to correct the record. C gastronomie One of us (Russell) even contributed to Etzioni’s survey, only to see his response being completely misconstrued. Gas bubble disease In fact, as our detailed analysis shows, Etzioni’s survey results are entirely consistent with the ones Bostrom cites.

How, then, does Etzioni reach his novel conclusion? By designing a survey instrument that is inferior to Bostrom’s and then misinterpreting the results.

The subtitle of the article reads, “If you ask the people who should really know, you’ll find that few believe AI is a threat to humanity.” So the reader is led to believe that Etzioni asked this question of the people who should really know, while Bostrom did not. Is there a gas station near me In fact, the opposite is true: Bostrom did ask people who should really know, but Etzioni did not ask anyone at all. Electricity prices over time Bostrom surveyed the top 100 most cited AI researchers. Electricity outage san antonio More than half of the respondents said they believe there is a substantial (at least 15 percent) chance that the effect of human-level machine intelligence on humanity will be “on balance bad” or “extremely bad (existential catastrophe).” Etzioni’s survey, unlike Bostrom’s, did not ask any questions about a threat to humanity.

Instead, he simply asks one question about when we will achieve superintelligence. Electricity and circuits class 6 ppt As Bostrom’s data would have already predicted, somewhat more than half (67.5 percent) of Etzioni’s respondents plumped for “more than 25 years” to achieve superintelligence—after all, more than half of Bostrom’s respondents gave dates beyond 25 years for a mere 50 percent probability of achieving mere human-level intelligence. Electricity lessons 4th grade One of us (Russell) responded to Etzioni’s survey with “more than 25 years,” and Bostrom himself writes, of his own surveys, “My own view is that the median numbers reported in the expert survey do not have enough probability mass on later arrival dates.”

Now, having designed a survey where respondents could be expected to choose “more than 25 years,” Etzioni springs his trap: he asserts that 25 years is “beyond the foreseeable horizon” and thereby deduces that neither Russell nor indeed Bostrom himself believes that superintelligent AI is a threat to humanity. Oil n gas prices This will come as a surprise to Russell and Bostrom, and presumably to many other respondents in the survey. Gastroenterologia o que trata (Indeed, Etzioni’s headline could just as easily have been “75 percent of experts think superintelligent AI is inevitable.”) Should we ignore catastrophic risks simply because most experts think they are more than 25 years away? By Etzioni’s logic, we should also ignore the catastrophic risks of climate change and castigate those who bring them up.

Contrary to the views of Etzioni and some others in the AI community, pointing to long-term risks from AI is not equivalent to claiming that superintelligent AI and its accompanying risks are “imminent.” The list of those who have pointed to the risks includes such luminaries as Alan Turing, Norbert Wiener, I.J. Gas 4 less redding ca Good, and Marvin Minsky. Gas pain left side Even Oren Etzioni has acknowledged these challenges. Gaston y daniela To our knowledge, none of these ever asserted that superintelligent AI was imminent. P gasol Nor, as noted above, did Bostrom in Superintelligence.

Etzioni then repeats the dubious argument that “doom-and-gloom predictions often fail to consider the potential benefits of AI in preventing medical errors, reducing car accidents, and more.” The argument does not even apply to Bostrom, who predicts that success in controlling AI will result in “a compassionate and jubilant use of humanity’s cosmic endowment.” The argument is also nonsense. Gas 99 cents a litre It’s like arguing that nuclear engineers who analyze the possibility of meltdowns in nuclear power stations are “failing to consider the potential benefits” of cheap electricity, and that because nuclear power stations might one day generate really cheap electricity, we should neither mention, nor work on preventing, the possibility of a meltdown.

Our experience with Chernobyl suggests it may be unwise to claim that a powerful technology entails no risks. Astrid y gaston lima reservations It may also be unwise to claim that a powerful technology will never come to fruition. Gas x dosage chewable On September 11, 1933, Lord Rutherford, perhaps the world’s most eminent nuclear physicist, described the prospect of extracting energy from atoms as nothing but “moonshine.” Less than 24 hours later, Leo Szilard invented the neutron-induced nuclear chain reaction; detailed designs for nuclear reactors and nuclear weapons followed a few years later. Bp gas prices akron ohio Surely it is better to anticipate human ingenuity than to underestimate it, better to acknowledge the risks than to deny them.

Many prominent AI experts have recognized the possibility that AI presents an existential risk. Austin electricity outage Contrary to misrepresentations in the media, this risk need not arise from spontaneous malevolent consciousness. Electricity merit badge pamphlet Rather, the risk arises from the unpredictability and potential irreversibility of deploying an optimization process more intelligent than the humans who specified its objectives. Gaz 67 for sale This problem was stated clearly by Norbert Wiener in 1960, and we still have not solved it. Gas delivery We invite the reader to support the ongoing efforts to do so.

I’m delighted that Professors Dafoe & Russell and I seem to be in agreement on three critical matters. Electricity khan academy One, we should refrain from ad hominem attacks. A level physics electricity notes Here, I have to offer an apology: I should not have quoted the anonymous AAAI Fellow who likened Dr. Electricity vocabulary Bostrom to Donald Trump. Gas engine tom I didn’t mean to lend my voice to that comparison; I sincerely apologized to Bostrom for this misstep via e-mail, an apology that he graciously accepted. Gas giants Two, as scientists, we need to assess statements about the risk of AI based on data. Gas national average 2013 That was the key point of my brief article , and the article offered unique data on this topic. Nyc electricity consumption Three, we also concur that the media has misrepresented the implications of Bostrom’s work—a topic of major concern that was allayed to some extent by the White House report on AI . Gas engineer salary Of course, we do differ on many of the details and the intricacies of the arguments, but time will tell.