Tech experts doubt ChatGPT and A.I. ‘hallucinations’ will disappear, saying it is not fixable.

Tech experts doubt ChatGPT and A.I. 'hallucinations' will disappear, saying it is not fixable.

The Challenges of Hallucination in AI Systems: A Journey Towards Truthful Generative AI

AI

Imagine a world where machines have the ability to generate new text, assisting in everything from psychotherapy to composing legal briefs. This revolutionary technology, known as generative AI, has captured the attention of businesses, organizations, and even high school students. However, there is a crucial problem that plagues these AI systems: hallucination.

Described as the act of making things up, confabulation, or simply hallucination, this challenge poses serious implications for any industry relying on generative AI to get work done. “I don’t think that there’s any model today that doesn’t suffer from some hallucination,” remarked Daniela Amodei, co-founder and president of Anthropic, the creator of the chatbot Claude 2. These AI models are primarily designed to predict the next word in a sentence, and as a result, there will inevitably be inaccuracies in their outputs.

Major developers of large language models, such as Anthropic and OpenAI (creator of the popular ChatGPT), acknowledge this problem and are actively working to enhance the truthfulness of their AI systems. However, the timeline for achieving this goal and whether these systems will ever be reliable enough to, for instance, offer medical advice, remain uncertain. Some experts, like linguistics professor Emily Bender, argue that the inherent mismatch between the capabilities of the technology and its proposed use cases makes complete elimination of hallucination unattainable.

The stakes are high when it comes to the reliability of generative AI technology. According to the McKinsey Global Institute, this technology has the potential to contribute trillions of dollars to the global economy. Engaging chatbots is just one aspect of this AI frenzy, which includes the generation of new images, videos, music, and computer code. Language forms an integral part of almost all these tools.

Even established entities like Google have started harnessing the power of generative AI. They are pitching a news-writing AI product to organizations where accuracy is crucial, such as news outlets. The Associated Press, in partnership with OpenAI, is also exploring the use of this technology to improve its AI systems by utilizing a portion of AP’s extensive text archive. The implications of hallucination, however, become more pronounced when the technology is applied outside of general text generation.

An example can be found in the work of computer scientist Ganesh Bagler. Bagler has been collaborating with India’s hotel management institutes to develop AI systems capable of inventing recipes for South Asian cuisines, like new variations of rice-based biryani. In this context, a single “hallucinated” ingredient could mean the difference between a delicious meal and an inedible one. This highlights the significance of addressing hallucination not only in text generation but across various industries where AI is employed.

During his visit to India, Sam Altman, the CEO of OpenAI, received direct feedback on the hallucination issue from Bagler. While Altman expressed optimism about resolving this problem, he estimated that it would take around a year and a half or two years. He identified the need to strike a balance between creativity and perfect accuracy, allowing the AI models to adapt and recognize the appropriate context for each requirement.

For some experts, including University of Washington linguist Bender, the improvements that can be made to language models will not be sufficient to eliminate hallucination completely. Bender explains that language models are essentially systems that predict the likelihood of different word sequences based on training data. They are designed to mimic existing forms of writing, such as legal contracts or television scripts, by selecting the most plausible next word. While they can be tuned to increase accuracy, they inevitably have failure modes that can be more difficult for a reader to detect due to their obscurity.

Interestingly, these errors are not seen as a major problem for marketing firms that rely on AI systems like Jasper AI for writing pitches. In fact, Shane Orlick, the president of Jasper AI, views hallucination as an added bonus. The AI-generated ideas and angles conceived by Jasper often surprise and impress their clients. Jasper collaborates with companies like OpenAI, Anthropic, Google, and Meta (Facebook’s parent company) to offer AI language models tailored to specific needs. Each model excels in different areas, addressing concerns such as accuracy or the security of proprietary information.

Orlick recognizes that fixing the hallucination problem won’t be easy. He believes that companies like Google, with their commitment to ensuring high standards of factual content, will invest considerable resources in finding solutions. While perfection may be elusive, the continuous refinement of AI models is expected to lead to significant improvements over time.

Amidst the challenges of hallucination, techno-optimists like Microsoft co-founder Bill Gates express optimism about the future of AI models. Gates believes that with time, these models can be taught to distinguish fact from fiction. He highlights promising work on this front, referring to a 2022 paper from OpenAI. However, even Altman, while promoting AI products for various applications, admits that he trusts the answers generated by ChatGPT the least, sparking laughter from the crowd during a visit to Bagler’s university.

In conclusion, the issue of hallucination in AI systems poses significant challenges for the reliability and accuracy of generative AI. While developers are striving to improve the truthfulness of their models, the inherent limitations of the technology might prevent complete eradication of hallucination. However, the advancements made so far are powering a new era of AI applications, with potential economic implications on a global scale. As the journey towards truthful generative AI continues, it remains a captivating and evolving phenomenon at the intersection of technology, language, and human ingenuity.

Please note that the links to external sources mentioned in the original content have not been retained in this rewrite.