Saturday, September 23, 2023
SCIENMAG: Latest Science and Health News
No Result
View All Result
  • Login
  • HOME PAGE
  • BIOLOGY
  • CHEMISTRY AND PHYSICS
  • MEDICINE
    • Cancer
    • Infectious Emerging Diseases
  • SPACE
  • TECHNOLOGY
  • CONTACT US
  • HOME PAGE
  • BIOLOGY
  • CHEMISTRY AND PHYSICS
  • MEDICINE
    • Cancer
    • Infectious Emerging Diseases
  • SPACE
  • TECHNOLOGY
  • CONTACT US
No Result
View All Result
Scienmag - Latest science news from science magazine
No Result
View All Result
Home SCIENCE NEWS Technology and Engineering

Verbal nonsense reveals limitations of AI chatbots

September 14, 2023
in Technology and Engineering
0
Share on FacebookShare on Twitter

NEW YORK – The era of artificial-intelligence chatbots that seem to understand and use language the way we humans do has begun. Under the hood, these chatbots use large language models, a particular kind of neural network. But a new study shows that large language models remain vulnerable to mistaking nonsense for natural language. To a team of researchers at Columbia University, it’s a flaw that might point toward ways to improve chatbot performance and help reveal how humans process language. 

Chatbot Nonsense Test

Credit: Columbia University’s Zuckerman Institute

NEW YORK – The era of artificial-intelligence chatbots that seem to understand and use language the way we humans do has begun. Under the hood, these chatbots use large language models, a particular kind of neural network. But a new study shows that large language models remain vulnerable to mistaking nonsense for natural language. To a team of researchers at Columbia University, it’s a flaw that might point toward ways to improve chatbot performance and help reveal how humans process language. 

 

In a paper published online today in Nature Machine Intelligence, the scientists describe how they challenged nine different language models with hundreds of pairs of sentences. For each pair, people who participated in the study picked which of the two sentences they thought was more natural, meaning that it was more likely to be read or heard in everyday life. The researchers then tested the models to see if they would rate each sentence pair the same way the humans had. 

 

In head-to-head tests, more sophisticated AIs based on what researchers refer to as transformer neural networks tended to perform better than simpler recurrent neural network models and statistical models that just tally the frequency of word pairs found on the internet or in online databases. But all the models made mistakes, sometimes choosing sentences that sound like nonsense to a human ear. 

 

“That some of the large language models perform as well as they do suggests that they capture something important that the simpler models are missing,” said Dr. Nikolaus Kriegeskorte, PhD, a principal investigator at Columbia’s Zuckerman Institute and a coauthor on the paper. “That even the best models we studied still can be fooled by nonsense sentences shows that their computations are missing something about the way humans process language.”

 

Consider the following sentence pair that both human participants and the AI’s assessed in the study:

 

That is the narrative we have been sold. 

This is the week you have been dying. 

 

People given these sentences in the study judged the first sentence as more likely to be encountered than the second. But according to BERT, one of the better models, the second sentence is more natural. GPT-2, perhaps the most widely known model, correctly identified the first sentence as more natural, matching the human judgments.

 

“Every model exhibited blind spots, labeling some sentences as meaningful that human participants thought were gibberish,” said senior author Christopher Baldassano, PhD, an assistant professor of psychology at Columbia. “That should give us pause about the extent to which we want AI systems making important decisions, at least for now.” 

 

The good but imperfect performance of many models is one of the study results that most intrigues Dr. Kriegeskorte. “Understanding why that gap exists and why some models outperform others can drive progress with language models,” he said. 

 

Another key question for the research team is whether the computations in AI chatbots can inspire new scientific questions and hypotheses that could guide neuroscientists toward a better understanding of human brains. Might the ways these chatbots work point to something about the circuitry of our brains?

 

Further analysis of the strengths and flaws of various chatbots and their underlying algorithms could help answer that question.

 

“Ultimately, we are interested in understanding how people think,” said Tal Golan, PhD, the paper’s corresponding author who this year segued from a postdoctoral position at Columbia’s Zuckerman Institute to set up his own lab at Ben-Gurion University of the Negev in Israel. “These AI tools are increasingly powerful but they process language differently from the way we do. Comparing their language understanding to ours gives us a new approach to thinking about how we think.”

 

###

 

The paper, “Testing the limits of natural language models for predicting human language judgements,” was published online in Nature Machine Intelligence on September 14, 2023. Its full list of authors includes Tal Golan, Matthew Siegelman,  Nikolaus Kriegeskorte and Christopher Baldassano.

 



Journal

Nature Machine Intelligence

DOI

10.1038/s42256-023-00718-1

Method of Research

Data/statistical analysis

Subject of Research

Not applicable

Article Title

Testing the limits of natural language models for predicting human language judgements

Article Publication Date

14-Sep-2023

COI Statement

The authors declare no competing interests.

Tags: chatbotsLimitationsnonsenserevealsVerbal
Share25Tweet16Share4ShareSendShare
  • blank

    Null results research now published by major behavioral medicine journal

    1034 shares
    Share 414 Tweet 259
  • New research reveals gut microbiota link to colitis: intestinal epithelial axin1 deficiency offers protective effects

    66 shares
    Share 26 Tweet 17
  • New findings on hair loss in men

    65 shares
    Share 26 Tweet 16
  • Fruit flies offer clues to how brains make reward-based decisions

    65 shares
    Share 26 Tweet 16
  • The potential of solar cars in the world

    64 shares
    Share 26 Tweet 16
  • Ochsner offers tuition assistance to aspiring nurses and doctors

    64 shares
    Share 26 Tweet 16
ADVERTISEMENT

About us

We bring you the latest science news from best research centers and universities around the world. Check our website.

Latest NEWS

Null results research now published by major behavioral medicine journal

Corning® launches Videodrop, revolutionizing real-time nanoparticle detection and analysis

Grant awarded to University of Louisville law professor will fund climate adaptation project

Subscribe to Blog via Email

Enter your email address to subscribe to this blog and receive notifications of new posts by email.

Join 208 other subscribers

© 2023 Scienmag- Science Magazine: Latest Science News.

No Result
View All Result
  • HOME PAGE
  • BIOLOGY
  • CHEMISTRY AND PHYSICS
  • MEDICINE
    • Cancer
    • Infectious Emerging Diseases
  • SPACE
  • TECHNOLOGY
  • CONTACT US

© 2023 Scienmag- Science Magazine: Latest Science News.

Welcome Back!

Login to your account below

Forgotten Password?

Retrieve your password

Please enter your username or email address to reset your password.

Log In