Futurologist states the rise of artificial intelligence (AI) will make most human better off over the next decade, but many have concerns about how advances in AI will impact what it means to be human, to be productive and exercise free will.
From SIRI to automated-driving cars, artificial intelligence is progressing fastly. When science fiction frequent portrays AI as robots with human-like characteristics, AI can cover anything from Google’s search algorithms to self-driving electric car.
Artificial intelligence recently is properly known as narrow AI, in that it is designed to do a narrow task (e.g. facial recognition or only internet searches or only driving a car). But, the long-term target of many researchers is to create common AI. While narrow AI may outperform people at whatever its specific task is, like playing chess or solving equations, AGI would surpass humans at nearly every cognitive task.
Digital life is add human capacities and disrupting eons-old human activities. Code-driven systems have spread to more than half of the world’s resident in ambient information and connectivity, providing previously unimagined opportunities and unprecedented threats. As emerging algorithm-driven artificial intelligence continues to spread, will human be better off than they are today?
WHY THE RECENT INTEREST IN AI SAFETY
Stephen Hawking, Steve Wozniak, Elon Musk, Bill Gates, and many other big names in science and technology have recently expressed concern in the media and through open letters about the risks posed by artificial intelligence, joined by many leading artificial intelligence researchers. Why is the subject suddenly in the headlines?
The idea that the quest for strong artificial intelligence would ultimately succeed was long consideration of as science fiction, centuries or more away. However, thanks to recent resolution, many AI milestones, which experts viewed as decades away merely 5 years ago, have now been reached, making many experts take seriously the probability of superintelligence in our lifetime. While some experts still guess that human-level artificial intelligence is centuries away, most AI researches at the 2015 Puerto Rico Conference guessed that it would happen before 2060. Since it may take decennary to finish the need safety research, it is prudent to begin it now.
Because AI has the potential to become more intelligent than any us(human), we have no certain way of predicting how it will behave. We can’t use past technological developments as much of a fundamental, because we’ve never build anything that has the ability to, wittingly or unwittingly, outsmart us. The best example of what we could face may be our own evolution. Human now manage the planet, not because human the strongest, fastest or biggest, but because human are the smartest. If we’re no longer the smartest, are we assured to remain in control?
FLI’s position is that our civilization will evolve as long as we win the race between the growing power of technology and the wisdom with which we control it. In the case of AI technology, FLI’s position is that the great way to win that race is not to impede the former, but to accelerate the latter, by supporting artificial intelligence safety research.