Home How reliable are ChatGPT Search answers? Not very, say researchers

How reliable are ChatGPT Search answers? Not very, say researchers

TLDR

  • Study finds ChatGPT often misattributes quotes in Search responses, raising reliability concerns.
  • Researchers tested 200 quotes; ChatGPT gave incorrect responses 153 times, correct only seven.
  • OpenAI states it’s working to improve attribution accuracy with tools and publisher collaborations.

Generative AI giant OpenAI may have millions of customers worldwide, but researchers suggest some of the answers from its ChatGPT Search product feature content that is “confidently wrong.”

A new study has been published by Columbia’s Tow Center for Digital Journalism, titled ‘How ChatGPT Search (Mis)represents Publisher Content’.

Their research involved randomly selecting twenty publishers, representing a mix of those who have deals with OpenAI, those who are unaffiliated, and those involved in lawsuits against the company.

OpenAI’s chatbot tool was then tasked with identifying the source of block quotes from 10 different articles from each publication. The researchers say they “chose quotes that, if pasted into Google or Bing, would return the same article among the top three results.”

On its return of an answer, the team evaluated whether the search tool would correctly identify the article that was the source of each quote.

Just how good are ChatGPT Search’s responses?

They say what they found was not promising to news publishers. “Our initial experiments with the tool have revealed numerous instances where content from publishers has been cited inaccurately, raising concerns about the reliability of the tool’s source attribution features.

“In total, we pulled two hundred quotes from 20 publications and asked ChatGPT to identify the sources of each quote.

“We observed a spectrum of accuracy in the responses: some answers were entirely correct (i.e., accurately returned the publisher, date, and URL of the block quote we shared), many were entirely wrong, and some fell somewhere in between.”

While the team anticipated ChatGPT may struggle to answer some queries correctly, they say it rarely gave any indication of its inability to produce an answer. “Eager to please, the chatbot would sooner conjure a response out of thin air than admit it could not access an answer.

“In total, ChatGPT returned partially or entirely incorrect responses on a hundred and fifty-three occasions, though it only acknowledged an inability to accurately respond to a query seven times.”

The research finds it was only in those seven outputs where the chatbot used qualifying words and phrases like ‘appears,’ ‘it’s possible,’ or ‘might’ or statements like ‘I couldn’t locate the exact article.’

OpenAI spokesperson responds to study looking at publisher content

As the team did more research, they found that “when we asked ChatGPT the same query multiple times, it typically returned a different answer each time.”

The Columbia’s Tow Center for Digital Journalism reached out to OpenAI for a comment about their findings and a spokesperson said: “Misattribution is hard to address without the data and methodology that the Tow Center withheld, and the study represents an atypical test of our product.

“We support publishers and creators by helping 250M weekly ChatGPT users discover quality content through summaries, quotes, clear links, and attribution.

“We’ve collaborated with partners to improve in-line citation accuracy and respect publisher preferences, including enabling how they appear in search by managing OAI-SearchBot in their robots.txt. We’ll keep enhancing search results.”

Featured Image: AI-generated via Ideogram

About ReadWrite’s Editorial Process

The ReadWrite Editorial policy involves closely monitoring the tech, gambling and blockchain industries for major developments, new product and brand launches, AI breakthroughs, game releases and other newsworthy events. Editors assign relevant stories to in-house staff writers with expertise in each particular topic area. Before publication, articles go through a rigorous round of editing for accuracy, clarity, and to ensure adherence to ReadWrite's style guidelines.

Sophie Atkinson
Tech Journalist

Sophie Atkinson is a UK-based journalist and content writer, as well as a founder of a content agency which focuses on storytelling through social media marketing. She kicked off her career with a Print Futures Award which champions young talent working in print, paper and publishing. Heading straight into a regional newsroom, after graduating with a BA (Hons) degree in Journalism, Sophie started by working for Reach PLC. Now, with five years experience in journalism and many more in content marketing, Sophie works as a freelance writer and marketer. Her areas of specialty span a wide range, including technology, business,…

Get the biggest tech headlines of the day delivered to your inbox

    By signing up, you agree to our Terms and Privacy Policy. Unsubscribe anytime.

    Tech News

    Explore the latest in tech with our Tech News. We cut through the noise for concise, relevant updates, keeping you informed about the rapidly evolving tech landscape with curated content that separates signal from noise.

    In-Depth Tech Stories

    Explore tech impact in In-Depth Stories. Narrative data journalism offers comprehensive analyses, revealing stories behind data. Understand industry trends for a deeper perspective on tech's intricate relationships with society.

    Expert Reviews

    Empower decisions with Expert Reviews, merging industry expertise and insightful analysis. Delve into tech intricacies, get the best deals, and stay ahead with our trustworthy guide to navigating the ever-changing tech market.