The LLM identifies and labels semantic parts of a search query, correcting misspellings simultaneously.
Benefit: This segmentation helps refine search location (e.g., understanding "sf" means San Francisco), improve business name matching, and apply relevant filters automatically. Spell correction ensures misspelled queries still yield good results. RAG helps distinguish business names from topics/locations.
The LLM generates a list of related phrases for a query. These phrases are then used to find and highlight relevant snippets in user reviews, helping users quickly see why a business matches their search.
"... came here for a pre-show dinner and it was perfect! Fast service and delicious food, just steps away from the theater..."
- User Review Snippet
Benefit: By generating semantically related phrases (not just keywords), the LLM helps surface highly relevant review snippets, even if the user's exact query words aren't present. This significantly improves the user's ability to quickly assess if a business meets their specific needs (like needing a quick dinner near a theater). RAG helps tailor phrases based on typical business categories associated with the query.
Define task, prototype with powerful LLM (e.g., GPT-4), test on head queries.
Create golden dataset, fine-tune smaller/cheaper LLM (e.g., GPT4o-mini), pre-compute for high-volume queries.
Serve pre-computed results (cache), optionally use real-time smaller model (e.g., BERT/T5/GPT4o-mini) for tail queries.