AI Language Models and the Truthfulness Challenge
Wed, Aug 2, 2023 10:57 AM on Latest,
Spending sufficient time with ChatGPT and other artificial intelligence chatbots can quickly reveal their tendency to produce false information. Referred to as hallucination, confabulation, or simply fabricating content, this issue poses a challenge for businesses, organizations, and even high school students relying on generative AI systems to compose documents and perform tasks. Some individuals use these systems for critical tasks, such as psychotherapy or researching and writing legal briefs. However, Daniela Amodei, co-founder and president of Anthropic, the creator of the chatbot Claude 2, admits that virtually all models suffer from some degree of hallucination. These models are primarily designed to predict the next word, and as a result, there will be inaccuracies in their predictions.
Major developers of large language models, including Anthropic, OpenAI (the maker of ChatGPT), and others, are actively working to enhance the truthfulness of their AI systems. Nonetheless, there are doubts about whether they will ever achieve sufficient accuracy, especially for critical tasks like dispensing medical advice. According to linguistics professor Emily Bender, the inherent mismatch between the technology and its proposed use cases makes the problem unfixable.
The reliability of generative AI technology carries significant importance, as the McKinsey Global Institute projects it to contribute trillions of dollars to the global economy. Chatbots represent only a portion of this trend, which also encompasses AI technology capable of generating images, videos, music, and computer code, with language components present in almost all these tools. For instance, Google is promoting a news-writing AI product to news organizations, where accuracy is of utmost importance.
Experts and researchers are exploring the potential of AI systems in various fields, such as computer scientist Ganesh Bagler's efforts to develop AI systems capable of inventing recipes for South Asian cuisines. However, the issue of hallucination remains concerning, particularly when it comes to tasks like recipe generation.
While some, like Sam Altman, the CEO of OpenAI, express optimism about mitigating the hallucination problem, experts like Emily Bender argue that improving these models will not be enough. Language models are designed to make things up, and even if they become more accurate, they will still have failure modes that might go unnoticed by readers.
For marketing firms using AI language models, hallucinations can be seen as advantageous, as they generate novel and creative ideas. Shane Orlick, the president of Jasper AI, emphasizes that his company collaborates with various partners to offer tailored AI language models based on their specific needs and concerns regarding accuracy and security.
Bill Gates and other techno-optimists believe that AI models can be taught to differentiate between fact and fiction over time. However, even Altman himself acknowledges that he trusts the answers from ChatGPT the least out of anyone.
In conclusion, the issue of hallucination in AI language models is a complex challenge, and while efforts are being made to improve their accuracy, achieving a perfect solution remains uncertain. The implications of AI's truthfulness have far-reaching consequences, and technological advancements will play a crucial role in addressing and refining this issue over time.
Source: Associated Press (AP)
