Scientists find ChatGPT is inaccurate when answering computer programming questions

Daily News
4 Min Read
Scientists find ChatGPT is inaccurate when answering computer programming questions

Editors’ notes

This article has been reviewed according to Science X’s editorial process and policies. Editors have highlighted the following attributes while ensuring the content’s credibility:

fact-checked

trusted source

proofread

by Bob Yirka , Tech Xplore

Quality of answers as rated by participants. Credit: Proceedings of the CHI Conference on Human Factors in Computing Systems (2024). DOI: 10.1145/3613904.3642596

A team of computer scientists at Purdue University has found that the popular LLM, ChatGPT, is wildly inaccurate when responding to computer programming questions. In their paper published as part of the Proceedings of the CHI Conference on Human Factors in Computing Systems, the group describes how they pulled questions from the StackOverflow website and posed them to ChatGPT and then measured its degree of accuracy when responding.

The team also presented their findings at the Conference on Human Factors in Computing Systems (CHI 2024) held May 11–16.

ChatGPT and other LLMs have been in the news a lot recently—since such apps have been made available to the general public, they have become very popular. Unfortunately, along with a treasure trove of useful information included in many of the responses given by such apps, there are a host of inaccuracies. Even more unfortunate is that it is not always clear when the apps are giving answers that are wrong.

In this new study, the team at Purdue noted that many programming have begun using LLMs to not only help write code for programming assignments, but to answer questions related to programming. As an example, a student could ask ChatGPT, what is the difference between a bubble sort and merge sort, or, more popularly, what is recursion?

To find out how accurate LLMs are in answering such questions, the research team focused their efforts on just one of them—ChatGPT. To find questions to use for testing the app, the researchers used questions freely available on the StackOverflow website—it is a site that has been built to help programmers learn more about programming by working with others in their field of interest. On one part of the site, users can post questions that will be answered by others who know the answers.

The research team used 517 questions found on the site and then measured how often ChatGPT gave the correct answer. Sadly, it was just 52% of the time. They also found the answers tended to be more verbose than would be the case when a human expert was asked the same question.

Alarmingly, the team found that user study participants preferred the answers given by ChatGPT 35% of the time. The researchers also found that the same users reading the answers given by ChatGPT quite often did not catch the mistakes that were made—they overlooked wrong answers 39% of the time.

More information: Samia Kabir et al, Is Stack Overflow Obsolete? An Empirical Study of the Characteristics of ChatGPT Answers to Stack Overflow Questions, Proceedings of the CHI Conference on Human Factors in Computing Systems (2024). DOI: 10.1145/3613904.3642596

© 2024 Science X Network

Citation: Scientists find ChatGPT is inaccurate when answering computer programming questions (2024, May 27) retrieved 27 May 2024 from https://techxplore.com/news/2024-05-scientists-chatgpt-inaccurate.html

This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.



Share This Article
Leave a comment