News — Language learning machines, such as ChatGPT, have become proficient in solving complex mathematical problems, passing difficult exams, and even offering advice for interpersonal conflicts. However, at what point does a helpful tool become a threat? 

Trust in AI is undermined because there is no science that predicts when its output goes from being informative and based on facts to producing material or even advice that is misleading, wrong, irrelevant or even dangerous. 

In a new study, George Washington University researchers explored when and why the output of large language models goes awry. , a professor of physics at the George Washington University, and a GW graduate student, Frank Yingjie Huo, developed a mathematical formula to pinpoint the moment at which the “Jekyll-and-Hyde tipping point” occurs. At the tipping point, AI’s attention has been stretched too thin and it starts pushing out misinformation and other negative content, Johnson says.

In the future, Johnson says the model may pave the way toward solutions which would help keep AI trustworthy and prevent this tipping point.

This paper provides a unique and concrete platform for discussions between the public, policymakers and companies about what might go wrong with AI in future personal, medical, or societal settings -- and what steps should be taken to mitigate the risks, Johnson says.

The study, was published as a white paper in arXiv. If you would like to schedule an interview with the researcher, please contact Claire Sabin, [email protected].

MEDIA CONTACT
Register for reporter access to contact details