Developed a RAG chatbot capable of answering queries from enterprise documents (PDF, DOCX). Integrated Ollama for local LLM inference, eliminating reliance on external APIs.
This project demonstrates the implementation of a Retrieval-Augmented Generation (RAG) chatbot capable of processing and answering queries from various enterprise document formats including PDF and DOCX files. The solution leverages Ollama for local LLM inference, eliminating dependencies on external APIs while maintaining high performance and data privacy.
FastAPI with Python for REST API endpoints
React.js for user interface
ChromaDB for vector storage and retrieval
Ollama for local language model inference
Docker Compose for multi-service orchestration