Google CEO admits he doesn't 'fully understand' how his AI works after it taught itself a new language and invented fake data to advance an idea



Google released Bard in March, an artificial intelligence tool touted as ChatGPT's rival. Just weeks into this public experiment, Bard has already defied expectations and ethical boundaries.

In an interview with CBS' "60 Minutes" that aired Sunday, Google CEO Sundar Pichai admitted that there is a degree of impenetrability regarding generative AI chatbots' reasoning.

"There is an aspect of this which we call ... a 'black box.' You know, you don't fully understand," said Pichai. "You can't quite tell why it said this or why it got wrong. We have some ideas, and our ability to understand this gets better over time. But that's where the state of the art is."

CBS' Scott Pelley asked, "You don't fully understand how it works and yet you've turned it loose on society?"

"Let me put it this way: I don't think we fully understand how a human mind works either," responded Pichai.

Despite citing ignorance on another subject as a rationale for blindly releasing new technology into the wild, Pichai was nevertheless willing to admit, "AI will impact everything."

Google describes Bard on its website as "a creative and helpful collaborator" that "can supercharge your imagination, boost your productivity, and help you bring your ideas to life—whether you want help planning the perfect birthday party and drafting the invitation, creating a pro & con list for a big decision, or understanding really complex topics simply."

While Pichai and other technologists have highlighted possible benefits of generative AI, Goldman Sachs noted in a March 26 report, "Generative AI could expose the equivalent of 300 million full-time jobs to automation."

In 2019, then-candidate Joe Biden told coal miners facing unemployment to "learn to code." In a twist of fate, the Goldman Sachs report indicated that coders and technologically savvy white-collar workers face replacement by Bard-like AI models at higher rates than those whose skills were only yesteryear denigrated by the president.

Legal, engineering, financial, sales, forestry, protective service, and education industries all reportedly face over 27% workforce exposure to automation.

In addition to losing hundreds of millions of jobs, truth may also be lost in the corresponding inhuman revolution.

"60 Minutes" reported that James Manyika, Google's senior vice president of technology and society, asked Bard about inflation. Within moments, the tool provided him with an essay on economics along with five recommended books, ostensibly as a means to bolster its claims. However, it soon became clear that none of the books were real. All of the titles were pure fictions.

Pelley confronted Pichai about the chatbot's apparent willingness to lie, which technologists reportedly refer to as "error with confidence" or "hallucinations."

For instance, according to Google, when prompted about how it works, Bard will often times lie or "hallucinate" about how it was trained or how it functions.

"Are you getting a lot of hallucinations?" asked Pelley.

"Yes, you know, which is expected. No one in the, in the field, has yet solved the hallucination problems. All models do have this as an issue," answered Pichai.

The Google CEO appeared uncertain when pressed on whether AI models' eagerness to bend the truth to suit their ends is a solvable problem, though noted with confidence, "We'll make progress."

\u201cOne AI program spoke in a foreign language it was never trained to know. This mysterious behavior, called emergent properties, has been happening \u2013 where AI unexpectedly teaches itself a new skill. https://t.co/v9enOVgpXT\u201d
— 60 Minutes (@60 Minutes) 1681687340

Bard is not just a talented liar. It's also an autodidact.

Manyika indicated Bard has evidenced staggering emergent properties.

Emergent properties are the attributes of a system that its constituent parts do not have on their own but arise when interacting collectively or in a wider whole.

Britannica offers a human memory as an example: "A memory that is stored in the human brain is an emergent property because it cannot be understood as a property of a single neuron or even many neurons considered one at a time. Rather, it is a collective property of a large number of neurons acting together."

Bard allegedly had no initial knowledge of or fluency in Bengali. However, need precipitated emergence.

"We discovered that with very few amounts of prompting in Bengali, it can now translate all of Bengali. So now, all of a sudden, we now have a research effort where we're now trying to get to a thousand languages," said Manyika.

These talented liars capable of amassing experience unprompted may soon express their competence in the world of flesh and bone — on factory floors, on soccer fields, and in other "human environments."

Raia Hadsell, vice president of research and robotics at Google's DeepMind, told "60 Minutes" that engineers helped teach their AI program how to emulate human movement in a soccer game. However, they prompted the self-learning program not to move like a human, but to learn how to score.

Accordingly, the AI, preoccupied with the ends, not the means, evolved its understanding of motion, discarding ineffective movements and optimizing its soccer moves in order to ultimately score more points.

"This is the type of research that can eventually lead to robots that can come out of the factories and work in other types of human environments. You know, think about mining, think about dangerous construction work or exploration or disaster recovery," said Hadsell.

The aforementioned Goldman Sachs report on jobs lost by automation did not appear to factor in the kind of self-learning robots Hadsell envisions marching out into the world.

Prior to the conquest of human environments by machines, there are plenty of threats already presented by these new technologies that may first need to be addressed.

Newsweek prompted Bard's competitor ChatGPT about risks that AI technology could pose, and it answered: "As AI becomes more advanced, it could be used to manipulate public opinion, spread propaganda, or launch cyber-attacks on critical infrastructure. AI-powered social media bots can be used to amplify certain messages or opinions, creating the illusion of popular support or opposition to a particular issue. AI algorithms can also be used to create and spread fake news or disinformation, which can influence public opinion and sway elections."

Like Blaze News? Bypass the censors, sign up for our newsletters, and get stories like this direct to your inbox. Sign up here!

Google releases Bard, a ChatGPT competitor



Google announced the rollout of Bard Tuesday, an artificial intelligence tool to rival ChatGPT.

"Today we’re starting to open access to Bard, an early experiment that lets you collaborate with generative AI," Google's product VP Sissie Hsiao and research VP Eli Collins wrote in a blog post partially generated by Bard.

The initial rollout is limited to "trusted testers," but residents of the United States and the United Kingdom were welcomed to sign up for a waiting list, Fox Business reported.

Bard is "powered by a research large language model," Google says, and will be updated over time. LLMs, the company says, are essentially prediction engines that act on one word at a time, when prompted. The more people use LLMs, the better they get a predicting the most helpful responses.

In Google's post, the company makes careful note of AI's limitations and risks.

"Because they learn from a wide range of information that reflects real-world biases and stereotypes, those sometimes show up in their outputs. And they can provide inaccurate, misleading or false information while presenting it confidently," the post says.

The post's authors included a humorous example of the technology's limitations when they used it to help create the post itself. When prompted with "give us a fun way Bard might wrap up this blog post," Bard responded, "And that's all for today, folks! Thanks for reading, and as always, remember to stay hydrated and eat your vegetables. Until next time, Bard out!"

Google's post says the company has built in "guardrails" to help ensure quality and safety. It also notes that the company is using human feedback and evaluation to help improve its systems.

Access to Bard for the U.S. and U.K. rolled out Tuesday. More countries and languages will be added over time, the company says. Those interested in test-driving the chatbot can sign up at bard.google.com.

Users need a personal Google account and must be aged 18 or over. According to Bard's FAQ page, the company will not use Bard conversations for advertising purposes at this time.

On signing up for Bard, would-be users may receive an automated reply thanking them and confirming inclusion on the waitlist.

"In the meantime, we asked Bard to write you a little poem while you wait," one such reply, received Tuesday afternoon, said.
"May your day be bright, your mood be light, and your heart filled with delight," read the included poem reportedly composed by Bard.
Like Blaze News? Bypass the censors, sign up for our newsletters, and get stories like this direct to your inbox. Sign up here!