After a Stanford survey of researchers showed more than a third believe that AI (artificial intelligence) could cause a "nuclear-level catastrophe," one of the directors of the California college's Institute for Human-Centered Artificial Intelligence said another key to that fear is deciphering what values system the human-inputted but self-determining bots will hold.
"I think that there is a little bit of overwhelming concern by a minority of people who are developing this technology. And one part of that is the concern that this technology will spin out of control and out of our hands," said Russell Wald, the institute's managing director for policy and society, said Monday on "The Story with Martha MacCallum."
"I think that is a probably fairly limited subset. But the bigger issue here I think we need to look at is who has a seat at this table -- and ensuring that we have a more diverse set of people at the table so that we can get out of this hysteria a little bit."
Wald said that at the current moment when entities like ChatGPT and others are developing hyper-intelligence, it is important to have a handle on whether the developer has humankind's best interests in mind.
STANFORD RESEARCHERS CREATE MINI WESTWORLD SIMULATION
At present, he said only China and the United States have entities with that ability, with the inference being the fact both countries have radically different foundations and values systems.
"Under that, the question is whose values are at the table? Is it Silicon Valley's and Beijing's or is it the rest of the country?" he said.
"And if we want the rest of the country at the table, we're going to have to provide a level of compute infrastructure to make it available for other stakeholders to be here and to have a more balanced and thoughtful view about this."
Anchor Martha MacCallum replied that Google executive Sundar Pichai reportedly spoke out in favor of AI entities reflecting American and Western values, showing that it is a legitimate concern how AI will respond in subjective terms to its objective outputs.
AI EXPERT ALARMED AFTER CHAT-GPT DEVISES PLAN TO ‘ESCAPE’
In a preview segment for Tucker Carlson's wide-ranging interview with Elon Musk airing Monday at 8 PM ET on Fox News, the Tesla and Twitter CEO affirmed the host's concern that some developers are teaching AI to "lie."
"That would be ironic," Musk replied. "But faith, the most ironic outcome, is most likely, it seems."
Wald said he doesn't believe AI is being trained to lie, but instead that there is "overconfidence" in the data systems it is being trained on – in that they will begin to invent context and create its own responses.
"That is a concern. And it's this emergent area that we're starting to study and have to understand a little bit more as to why that's happening," he said. "And a good part of that is the data."
"The other part of this is that we need to ensure that the right values are at this and that we have a diverse set of views so that we don't have this just cultural-relativist side, but we actually have a full diversity of views that are coming through on some of this technology."