Back to Projects
Professional ProjectAI/ML

Company-Wide RAG-based Chat Assistant

Timeline: 2022
Role: AI Solution Architect

Overview

Architected and integrated self-hosted enterprise RAG system using Langchain (early adoption), vector databases, and agent frameworks to provide intelligent access to company knowledge base.

Challenge

Support team received many requests requiring developer knowledge. Information scattered across multiple platforms. No unified way to access company documentation. Early-stage embedding models were error-prone.

Solution & Approach

Designed and integrated RAG system using Langchain immediately after its release with Azure OpenAI Service. Integrated multiple company platforms into Weaviate vector database. Implemented multi-modal embeddings (text and image) for comprehensive semantic search. Architected self-hosted solution for data privacy and compliance. Evaluated and migrated between embedding models: started with gte-large, migrated to qwen3-embedding for improved performance.

Outcome & Impact

Reduced support requests to Development team by 50%. Semantic search dramatically outperformed traditional Confluence keyword search, with employees providing feedback that the chatbot found far more relevant information than Confluence's search function. Deployed internally and actively used across company. Enabled self-service for developer-level questions. First production LLM application integrated at company.

Technologies Used

LangchainAzure OpenAI ServiceWeaviateRAGPythongte-largeqwen3-embeddingMulti-modal embeddingsAgent-based architecture

Key Highlights

  • Reduced support requests to Development team by 50%
  • Early adopter of Langchain (v0.x)
  • Implemented multi-modal embeddings (text and image)
  • Self-hosted solution for data privacy
  • First production LLM application at company
View All Projects