Eduardo Ordaxโ€™s Post

View profile for Eduardo Ordax, graphic

๐Ÿค– Generative AI Lead @ AWS โ˜๏ธ | Startup Advisor | Public Speaker

๐Ÿš€ ๐—”๐—ฑ๐—ฎ๐—ฝ๐˜๐—ถ๐˜ƒ๐—ฒ ๐—ฎ๐—ป๐—ฑ ๐—ฆ๐—ฒ๐—น๐—ณ-๐—ฅ๐—ฒ๐—ณ๐—น๐—ฒ๐—ฐ๐˜๐—ถ๐˜ƒ๐—ฒ ๐—ฅ๐—ฒ๐˜๐—ฟ๐—ถ๐—ฒ๐˜ƒ๐—ฎ๐—น ๐—”๐˜‚๐—ด๐—บ๐—ฒ๐—ป๐˜๐—ฒ๐—ฑ ๐—š๐—ฒ๐—ป๐—ฒ๐—ฟ๐—ฎ๐˜๐—ถ๐—ผ๐—ป (๐—ฅ๐—”๐—š) Ever wondered how to make your LLMs smarter and more reliable? Imagine a system that not only retrieves information but also corrects itself to provide accurate responses. Welcome to the world of RAG with self-correction! ๐Ÿค–๐Ÿ“š ๐—›๐—ผ๐˜„ ๐—ถ๐˜ ๐˜„๐—ผ๐—ฟ๐—ธ๐˜€ ๐Ÿ‘‡ 1๏ธโƒฃ ๐—ค๐˜‚๐—ฒ๐˜€๐˜๐—ถ๐—ผ๐—ป ๐—ฅ๐—ผ๐˜‚๐˜๐—ถ๐—ป๐—ด ๐—ก๐—ผ๐—ฑ๐—ฒ: Routes questions to either document retrieval or web search based on relevance. 2๏ธโƒฃ ๐—ฅ๐—ฒ๐˜๐—ฟ๐—ถ๐—ฒ๐˜ƒ๐—ฒ๐—ฟ ๐—ก๐—ผ๐—ฑ๐—ฒ: Transforms questions into embeddings and retrieves relevant documents from the vector store. 3๏ธโƒฃ ๐—š๐—ฟ๐—ฎ๐—ฑ๐—ถ๐—ป๐—ด ๐——๐—ผ๐—ฐ๐˜‚๐—บ๐—ฒ๐—ป๐˜๐˜€ ๐—ก๐—ผ๐—ฑ๐—ฒ: LLM grades the retrieved documents for relevance. 4๏ธโƒฃ ๐—š๐—ฒ๐—ป๐—ฒ๐—ฟ๐—ฎ๐˜๐—ถ๐—ป๐—ด ๐—”๐—ป๐˜€๐˜„๐—ฒ๐—ฟ๐˜€ ๐—ก๐—ผ๐—ฑ๐—ฒ: LLM generates answers if the documents are deemed sufficient. 5๏ธโƒฃ ๐—ฆ๐—ฒ๐—น๐—ณ-๐—–๐—ผ๐—ฟ๐—ฟ๐—ฒ๐—ฐ๐˜๐—ถ๐—ผ๐—ป ๐— ๐—ฒ๐—ฐ๐—ต๐—ฎ๐—ป๐—ถ๐˜€๐—บ: ๐Ÿ”ธ ๐—ก๐—ผ ๐—›๐—ฎ๐—น๐—น๐˜‚๐—ฐ๐—ถ๐—ป๐—ฎ๐˜๐—ถ๐—ผ๐—ป๐˜€: Accurate answers proceed to the next check. ๐Ÿ”ธ ๐—›๐—ฎ๐—น๐—น๐˜‚๐—ฐ๐—ถ๐—ป๐—ฎ๐˜๐—ถ๐—ผ๐—ป๐˜€: Inaccurate answers return to the Generation Node for refinement. 6๏ธโƒฃ ๐—ฉ๐—ฎ๐—น๐—ถ๐—ฑ๐—ฎ๐˜๐—ถ๐—ป๐—ด ๐˜๐—ต๐—ฒ ๐—™๐—ถ๐—ป๐—ฎ๐—น ๐—”๐—ป๐˜€๐˜„๐—ฒ๐—ฟ: ๐Ÿ”ธ ๐—ฌ๐—ฒ๐˜€: Accurate answers are provided to the user. ๐Ÿ”ธ ๐—ก๐—ผ: The system performs a web search for additional information. 7๏ธโƒฃ ๐—ช๐—ฒ๐—ฏ ๐—ฆ๐—ฒ๐—ฎ๐—ฟ๐—ฐ๐—ต ๐—œ๐—ป๐˜๐—ฒ๐—ด๐—ฟ๐—ฎ๐˜๐—ถ๐—ผ๐—ป: Conducts a web search when necessary for additional data or to correct hallucinations. This framework empowers LLMs to dynamically retrieve diverse data sources, refine responses autonomously, and ensure reliability through self-correction mechanisms. ๐Ÿ“Ž Credits to medium post below:https://lnkd.in/dSucDu9Y #rag #ai #llmops #genai

  • No alternative text description for this image
Arsenii Kvachan

Artificial Intelligence Engineer | NLP, DL, ML

1w

How do you detect hallucinations?

Sunil Mehta

CEO at Graph & NimbleCat

1w

How many round trips to a LLM are required to get the benefit of this on average?

Ali Pala

Test Automation Coach | AI Apps Enthusiast | GenAI | LLMs

1w

Also, do such number of steps make the pipeline less efficient or slower? Read and article yesterday about multi agent solutions. Would those be a solution? What do you think? For example Autogen or crewAI.

Ali Pala

Test Automation Coach | AI Apps Enthusiast | GenAI | LLMs

1w

Great post, thank you. Iโ€™m wondering how the model ensures about retrieved data from web search is reliable to correct hallucinations? Might be a rare case but not impossible.

Sandi Bezjak

AI - QUANTUM COMPUTER - NANO TECH - AR - VR - BIO TECH or Everything of everything | Information Technology Analyst

1w

Here's something I asked in perplexity.ai: adoptive self reflecting retrieval augmented generation self reflecting graphs and all other self reflecting options in anything what you recommend and new never tried options and you are a Grandmaster All-knowing Genius https://www.perplexity.ai/search/adoptive-self-reflecting-retri-Pod2xSijQN.MOCbIyFe00g

Carlos Escapa

Global Business Developer of Data & AI Solutions, guest lecturer

1w

how does the "self correction" mechanism detect hallucinations?

Stanislav Dalence, PMP, PSM, LBBP

Top Artificial Intelligence (AI) Voice - Director de Proyecto Senior in NICE CXone

1w

my question would be how to get the "is Hallucination?" question answered via self reflection in order to further process your workflow without human intervention

Like
Reply
Dennis D.

Hand-crafted AI solutions for established companies

1w

How do you optimize latency to ensure user experience over that many processing steps between user intent and final answer generation? And how is hallucination measured quantitatively?

Like
Reply
Peter Jeitschko

Co-Founder CRO at JetHire.ai | Advanced Generative AI for HR | Find Talent with AI

1w

I wonder how GraphRAG fits into this picture

Marcelo Grebois

โ˜ฐ Cloud & Software Architect โ˜ฐ MLOps โ˜ฐ AIOps โ˜ฐ Helping companies scale their platforms to an enterprise grade level

1w

The RAG model enhances LLM reliability by dynamically refining responses with self-correction mechanisms. Experience the power of innovative AI Eduardo Ordax

See more comments

To view or add a comment, sign in

Explore topics