Verbal nonsense reveals limitations of AI chatbots

The era of artificial intelligence (AI) chatbots that appear to understand and use language in a human-like manner has dawned. These chatbots rely on large language models, a type of neural network. However, a recent study has revealed a vulnerability in these large language models, as they can sometimes mistake nonsense for natural language. Researchers at Columbia University see this flaw as an opportunity to enhance chatbot performance and gain insights into how humans process language.

In their paper published in Nature Machine Intelligence, the scientists describe how they conducted experiments using nine different language models. They presented hundreds of pairs of sentences to human participants and asked them to select the sentence they believed sounded more natural, i.e., the one more likely to be encountered in everyday communication. The researchers then evaluated whether the AI models would provide the same judgments as the human participants.

In head-to-head comparisons, the more advanced AI models based on transformer neural networks generally outperformed simpler models, such as recurrent neural networks and statistical models that rely on word pair frequencies from the internet or online databases. However, all models exhibited errors, occasionally selecting sentences that sounded like gibberish to humans.

Dr. Nikolaus Kriegeskorte, a principal investigator at Columbia’s Zuckerman Institute and a coauthor of the paper, noted, “That some of the large language models perform as well as they do suggests that they capture something important that the simpler models are missing. That even the best models we studied still can be fooled by nonsense sentences shows that their computations are missing something about the way humans process language.”

For example, consider the following sentence pair:

  1. That is the narrative we have been sold.
  2. This is the week you have been dying.

Human participants in the study judged the first sentence as more natural. However, BERT, one of the advanced models, rated the second sentence as more natural, while GPT-2, another widely known model, correctly identified the first sentence as more natural, aligning with human judgments.

Christopher Baldassano, an assistant professor of psychology at Columbia and the senior author of the study, emphasized that all models had blind spots and labeled some sentences as meaningful when human participants considered them gibberish. He cautioned against relying too heavily on AI systems for important decisions, at least in their current state.

One of the intriguing findings of the study is the good yet imperfect performance of many models. Dr. Kriegeskorte emphasized the importance of understanding why these gaps exist and why certain models outperform others, as this knowledge can drive progress in language models.

The researchers are also curious about whether the computations in AI chatbots can inspire new scientific questions and hypotheses, potentially guiding neuroscientists toward a better understanding of human brain function. Analyzing the strengths and weaknesses of various chatbots and their underlying algorithms may contribute to answering this question.

Tal Golan, the paper’s corresponding author, who recently established his own lab at Ben-Gurion University of the Negev in Israel, highlighted the interest in understanding how people think and the unique processing of language by AI tools, offering a fresh perspective on human cognition.

Posted in

Aihub Team

Leave a Comment





Is AI electricity or the telephone?

Is AI electricity or the telephone?

Introducing Superalignment

Introducing Superalignment

GPT-4 API general availability and deprecation of older models in the Completions API

GPT-4 API general availability and deprecation of older models in the Completions API

Democratic inputs to AI

Democratic inputs to AI

DALL-E 2 Chimera prompts

DALL-E 2 Chimera prompts

Can AI predict the future?

Can AI predict the future?

Bing is sadly too desperate to make AI work

Bing is sadly too desperate to make AI work

AI progress is scaring people

AI progress is scaring people

AI in the modeling industry

AI in the modeling industry

AI Driven Testing

AI Driven Testing

AI as Co-Creator of Test Design

AI as Co-Creator of Test Design

 The Good, The Bad, & The Hallucinatory – How AI can help and hurt secure development

 The Good, The Bad, & The Hallucinatory – How AI can help and hurt secure development

The CX Paradigm Shift: Exploring Generative AI’s Impact on Customer Experience

The CX Paradigm Shift: Exploring Generative AI’s Impact on Customer Experience

Edge Computing Expo Europe, 26-27 September 2023

Edge Computing Expo Europe, 26-27 September 2023

Digital Transformation Week Europe | 26-27 September 2023

Digital Transformation Week Europe | 26-27 September 2023

The Security of Artificial Intelligence

The Security of Artificial Intelligence

AI Combined with Automation is the Perfect Marriage for Scalable, Intelligent Operations

AI Combined with Automation is the Perfect Marriage for Scalable, Intelligent Operations

AI and Phishing: What’s the Risk to Your Organization?

AI and Phishing: What’s the Risk to Your Organization?

Why Claude AI is your new go-to for complex tasks

Why Claude AI is your new go-to for complex tasks

The Smart Home Jury Is Still Out on Matter, AI Could Help

The Smart Home Jury Is Still Out on Matter, AI Could Help

Explore Jasper AI, a writing tool that makes creators’ lives easier

Explore Jasper AI, a writing tool that makes creators’ lives easier

Enjoy the journey while your business runs on autopilot

Enjoy the journey while your business runs on autopilot

ChatGPT failed to get service status: Fixes and alternatives to try

ChatGPT failed to get service status: Fixes and alternatives to try

ChatGPT Down? OpenAI Chatbot ChatGPT Reportedly Hit by Global Outage, Users Lodge Complaints on Twitter

ChatGPT Down? OpenAI Chatbot ChatGPT Reportedly Hit by Global Outage, Users Lodge Complaints on Twitter

Blue Chip Ads Feeding Unreliable AI-Generated News Websites

Blue Chip Ads Feeding Unreliable AI-Generated News Websites

Social media algorithms are still failing to counter misleading content

Social media algorithms are still failing to counter misleading content

Rishabh Mehrotra, research lead, Spotify: Multi-stakeholder thinking with AI

Rishabh Mehrotra, research lead, Spotify: Multi-stakeholder thinking with AI

Researchers from Microsoft and global leading universities study the ‘offensive AI’ threat

Researchers from Microsoft and global leading universities study the ‘offensive AI’ threat

GTC 2021: Nvidia debuts accelerated computing libraries, partners with Google, IBM, and others to speed up quantum research

GTC 2021: Nvidia debuts accelerated computing libraries, partners with Google, IBM, and others to speed up quantum research

Facebook is developing a news-summarising AI called TL;DR

Facebook is developing a news-summarising AI called TL;DR