Have the risks of artificial intelligence risen to the point where more regulation is needed? Cognitive scientist Gary Marcus argues that the federal government — or maybe even international agencies — will need to step in.
The Food and Drug Administration or the Federal Aviation Administration could provide a model, Marcus said last week during a fireside chat with Seattle science-fiction author Ted Chiang at Town Hall Seattle.
“I think we would like to have something like an FDA-like approval process if somebody introduces a new form of AI that has considerable risks,” Marcus said. “There should be some way of regulating that and saying, ‘Hey, what are the costs? What are the benefits? Do the benefits to society really outweigh the costs?’”
