AskObserve
Production LLM · Conversational Analytics
- accuracy
- +35%
- latency
- −20%
- search
- <100ms
Spring-based microservice that puts an LLM in the loop for 25+ business units, with an intelligent selector that routes between Snowflake structured analytics and LLM/NLP sources.
- Built dual-source ML router that picks between structured analytics and LLM/NLP based on query shape
- Engineered sub-100ms semantic history search on Elasticsearch + MongoDB Atlas with compound indexes and fuzzy matching
- Designed tag-based filter pipeline (hierarchical teams, date-range merging, dynamic enrichments) translating UI selections into ML format
- AOP-based monitoring + cache-driven filter pipeline for low-latency reads