Cosmin Negruseri’s Post

View profile for Cosmin Negruseri, graphic

cofounder at stealth startup

Sharing something that may be obvious to people working in the search space but maybe a bit less obvious people new to genai, search or RAG. Traditional Search is getting updated in the post ChatGPT Era and is getting democratized. Traditional systems involve many steps: - Crawling, Indexing, Filtering, Deduplication - Query Understanding - User Understanding/Personalization - Document Understanding - Multiple Candidate Generators - Multiple Ranking Layers - Business Logic Integration These steps create a list of 10 blue links. In the post-ChatGPT world, LLMs can improve or replace many search components: - Better Document Understanding: Use LLMs to summarize and extract information from documents. - User Understanding: Use LLMs to understand what users like and engage with. - Ranking: Let LLMs rank answers based on relevance and quality. - Quality Rating and evaluation: Use LLMs to rate the quality of answers, reducing the need for manual raters. - Answer Generation (or rather Retrieval Augmented Generation): Provide direct answers with references, not just blue links and the ability to continue chatting while keeping the context. Think of it as having a smart high schooler to improve or replace any part of the old search system. This intelligence is democratized, building a search system is no longer only for big companies with lots of resources and having historical user actions.

  • No alternative text description for this image
John Milinovich

Head of GenAI Product at Canva

1w

While I think this will be true in a few years, we’re not there yet for doing inference at query time (ie query understanding and ranking) due to latency issues. Search p99 needs to be measured in milliseconds whereas LLM p99 can still be 10s+. I’m optimistic that fine tuned models and SLMs will help us get there soon-ish though.

Cătălin Moraru

Software Engineer at Google

1w

But won’t big companies have the best, most complex GenAI? Maybe it won’t be the same companies, but still does not seem to be a democracy.

Like
Reply
Robb Beal

Digital Product and Experience Leader | Startups: $75M Total Investment and 3 Exits

1w

For a user-centric perspective, see also: https://dl.acm.org/doi/pdf/10.1145/3649468

  • No alternative text description for this image
Like
Reply

The cost curve is still an issue for such an LLM-centric search stack to fully manifest at scale. I do expect that to happen as smaller models like Gemma, Phi and Mistral7B become smarter and serving infrastructure is further optimized.

See more comments

To view or add a comment, sign in

Explore topics