Select Page

akinkoyeoyindamola EVERY NEW TECHNOLOGY BRINGS its own nightmare sce-narios….EVERY NEW TECHNOLOGY BRINGS its own nightmare sce-narios. Artificial intelligence (AI) and robotics are no excep-tions. Indeed, the word “robot” was coined for a 1920 play that dramatized just such a doomsday for humanity. Recently, an open letter about the future of Al, signed by a number of high-profile scientists and entrepreneurs, spurred a new round of harrowing headlines like “Top Scientists Have an Ominous Warning about Artificial Intelligence,” and “Artificial Intelligence Experts Sign Open Letter to Protect Mankind from Machines.” The implication is that the machines will one day displace humanity. Let’s get one thing straight: a world in which humans are enslaved or destroyed by super intelligent machines of our own creation is purely science fiction. Like every other technology, Al has risks and benefits, but we cannot let fear dominate the conversation or guide Al research. Nevertheless, the idea of dramatically changing the Al research agenda to focus on Al “safety” is the primary message of a group calling itself the Future of Life Institute (FLI). FLI includes a handful of deep thinkers and public figures such as Elon Musk and Stephen Hawking and worries about the day in which humanity is steamrolled by powerful programs run amuck. As eloquently described in the book Super intelligence:Paths, Dangers, Strategies by FLI advisory board member and Oxford-based philosopher Nick Bistros, the plot unfolds in three parts. In the first part-roughly where we are now. computational power and intelligent software develops at an increasing pace through the toil of scientists and engineers.
Next, a breakthrough is made: programs are created that possess intelligence on par with humans. These programs, running on increasingly fast computers, improve themselves extremely rapidly, resulting in a runaway “intelligence explo-sion.” In the third and final act, a singular super-intelligence takes hold-outsmarting, outmaneuvering, and ultimately outcompeting the entirety of humanity and perhaps life itself. Let’s take a closer look at this apocalyptic storyline. Of the s three parts, the first is indeed happening now and Bostrom provides cogent and illuminating glimpses into current and near-future technology. The third part is a philosophical romp exploring the consequences of super smart machines. It’s that second part–the intelligence explosion -that demonstrably violates what we know of computer science and natural intelligence The notion of the intelligence explosion arises from Moore’s Law, the observation that the speed of computers has been increasing exponentially since the 1950s. Project this trend forward and we’ll see computers with the computational power of the entire human race within the next few decades. It’s a leap to go from this idea to unchecked growth of machine intelligence, however.
First, ingenuity is not the sole bottleneck to developing faster computers. The machines need to actually be built, which requires real-world resources. Indeed, Moore’s law comes with exponentially increasing production costs as well mass production of precision electronics does not come cheap. Further, there are fundamental physical laws quantum limits that bound how quickly a transistor can do its work. Non-silicon technologies may overcome those limits, but such devices remain highly speculative. In addition to physical laws, we know a lot about the fundamental nature of computation and its limits. For example, some computational puzzles, like figuring out how to factor a number and thereby crack online cryptography schemes, are generally believed to be unsolvable by any fast program. They are part of a class of mathematically defined problems that are “NP-complete,” meaning that they are exactly as hard as any problem that can be solved non-deterministically (N) in polynomial time (P), and they have resisted any attempt at scalable solution. As it turns out, most computational problems that we associate with human intelligence are known to be in this class. Wait a second, you might say. How does the human mind manage to solve mathematical problems that computer scientists believe can’t be solved? We don’t. By and large, we cheat. We build a cartoonish mental model of the elements of the world that we’re interested in and then probe the behaviour of this invented mini world. There’s a trade-off between completeness and tractability in these imagined microcosms. Our ability to propose and ponder and project credible futures comes at the cost of accuracy. Even allowing for the possibility of the existence of considerably faster computers than we have today, it is a logical impossibility that these computers would be able to accurately simulate reality faster than reality itself.
Countering the Anti-Al Cause In the face of general skepticism in the Al and computer science 10 communities about the possibility of an intelligence explosion, FLI still wants to win support for its cause. The group’s letter calls for increased attention to maximizing the societal benefits of developing Al. Many of my esteemed colleagues signed the letter to show their support for the importance of avoiding potential pitfalls of the technology. But a few key phrases in the letter such as “our Al systems must do what we want them to do” (Tegmark) are taken by the press as an admission that Al researchers believe they might be creating something that cannot be controlled. It also implies that Al researchers are asleep at the wheel, oblivious to the ominous possibilities, which is simply untrue.
To be clear, there are indeed concerns about the near-term future of Al- algorithmic traders crashing the economy, or sensitive power grids overreacting to fluctuations and shutting down electricity for large swaths of the population. There’s also a concern that systemic biases within academia and industry prevent underrepresented minorities from participating and helping to steer the growth of information technology. These worries should play a central role in the development and deployment of new ideas. But dread predictions of computers suddenly waking up and turning on us are simply not realistic.
I welcome an open discussion about how Al can be made robust and beneficial, and how we can engineer intelligent machines and systems that make society better. But, let’s please keep the discussion firmly within the realm of reason and leave the robot uprisings to Hollywoxxl screenwriters.

Summarize the article ” Rise of the machines” is not a likely future by Michael Littman Arts & HumanitiesEnglish