AI's Latest and Greatest Videos

Good News: A.I. Is Getting Cheaper. That’s Also Bad News.

In time, putting these pieces together — researchers call them dual-use technologies — will become increasingly easy and inexpensive. How hard would it be to make a similar but dangerous device?

“This stuff is getting more available in every sense,” said one of Skydio’s founders, Adam Bry. These same technologies are bringing a new level of autonomy to cars, warehouse robots, security cameras and a wide range of internet services.

But at times, new A.I. systems also exhibit strange and unexpected behavior because the way they learn from large amounts of data is not entirely understood. That makes them vulnerable to manipulation; today’s computer vision algorithms, for example, can be fooled into seeing things that are not there.

“This becomes a problem as these systems are widely deployed,” said Miles Brundage, a research fellow at the University of Oxford’s Future of Humanity Institute and one of the report’s primary authors. “It is something the community needs to get ahead of.”

Advances in artificial intelligence have made it easier to fake audio and video by grafting a person’s face onto another person’s body. Video by derpfakes

The report warns against the misuse of drones and other autonomous robots. But there may be bigger concerns in less obvious places, said Paul Scharre, another author of the report, who had helped set policy involving autonomous systems and emerging weapons technologies at the Defense Department and is now a senior fellow at the Center for a New American Security.

“Drones have really captured the imagination,” he said. “But what is harder to anticipate — and wrap our heads around — is all the less tangible ways that A.I. is being integrated into our lives.”

The rapid evolution of A.I. is creating new security holes. If a computer-vision system can be fooled into seeing things that are not there, for example, miscreants can circumvent security cameras or compromise a driverless car.

Researchers are also developing A.I. systems that can find and exploit security holes in all sorts of other systems, Mr. Scharre said. These systems can be used for both defense and offense.

Automated techniques will make it easier to carry out attacks that now require extensive human labor, including “spear phishing,” which involves gathering and exploiting personal data of victims. In the years to come, the report said, machines will be more adept at collecting and deploying this data on their own.

A.I. systems are increasingly adept at generating believable audio and video on their own. This will accelerate the progress of virtual reality, online games and movie animation. It will also make it easier for bad actors to spread misinformation online, the report said.

This is already beginning to happen through a technology called “Deepfakes,” which provides a simple way of grafting anyone’s head onto a pornographic video — or put words into the mouth of the president.

Some believe concerns over the progress of A.I. are overblown. Alex Dalyac, chief executive and co-founder of a computer vision start-up called Tractable, acknowledged that machine learning will soon produce fake audio and video that humans cannot distinguish from the real thing. But he believes other systems will also get better at identifying misinformation. Ultimately, he said, these systems will win the day.

To others, that sounds like an endless cat-and-mouse game between A.I. systems trying to create the fake content and those trying to identify it.

“We need to assume that there will be advances on both sides,” Mr. Scharre said.

Continue reading the main story

By CADE METZ

https://www.nytimes.com/2018/02/20/technology/artificial-intelligence-risks.html

Source link

Similar Posts

WP2Social Auto Publish Powered By : XYZScripts.com