Dusty Chadwick’s Post

View profile for Dusty Chadwick, graphic

VP of Engineering / Head of AI @ Voze | Leadership in AI (AI Utah 100 Honoree), Cloud Architecture and Data Engineering

Hallucinations in LLMs are inevitable, no seriously! Josh Fritzsche and I have worked hard at Voze to find a way to tune, prompt and even plead 🙏 with LLMs to prevent them from the occasional 🍄 hallucination. It seemed like the better we have done the harder the edge cases have been to solve and eventually we realized that they tend to frequently happen when LLMs predict or make anticipations on "nothing". I remember at one point nearly going insane with frustration because when LLMs had a hallucination, they were EPIC! 😱 Then we pushed through it and benefited for the effort! How did we solve this common problem? We didn't (Not completely), we got better at identifying the issues as they happened. Once we identified in the data we were able to start looking for patterns. The same catalysts that cause it to happen once if left unchanged caused it over and over again. Our solution is astoundingly simple. Once you know that it's likely to happen..... let it! Accept it will 🍄, but also be preemptive in not focusing on it and use alternative data sources, generators or providers. Change the conditions that cause it but continue to monitor it as it happens. As Dan Caffee is fond of saying to me on a regular bases. "Dusty we need the system to be adaptive and provide mechanisms to learn from it's past." Well he is right! All we needed was data and the people that contribute to it's quality. Huge props 👏 to strong support from Kathy and Janelles audit teams. We were armed with a knowledge and a desire to learn from past mistakes. Armed with confidence that we could correct these problems we could focus on the data to prevent their impact going forward. When doing millions of LLM (Generative AI) calls regularly Josh and I have learned that hallucinations will happen. We also know exactly how we will handle and prevent it going forward! Failure is not doing anything and accepting defeat.

Hallucination is Inevitable: An Innate Limitation of Large Language Models

Hallucination is Inevitable: An Innate Limitation of Large Language Models

arxiv.org

Ber Zoidberg

Building web apps and data pipelines at-scale

1mo

The reality is that we approach LLMs wrong, and I think we do this because we don’t understand the complex iterative process of human cognition. When you think about something, even something as simple as recalling information you have stored in your own head, most people go through a series of steps to self-validate that information before just spewing it out. LLMs don’t have this luxury — the functions we use for training don’t enforce this, the models don’t seem to naturally develop this within their networks, and we are so obsessed with simple single-iteration IO against a single universal model that we completely miss that human intelligence comes from internal iteration and cross-communication across multiple specialized neural networks. The future of AI, and the solution to hallucination, is engineering a system of networks working together to solve problems, answer questions, etc. Beyond what we currently attempt with MoE models.

To view or add a comment, sign in

Explore topics