Example: PDF Chatbot📚#
Description:
This example showcases how to build a PDF chatbot with local LLM and Embedding models
Used Technology:
@ Xinference as a LLM model hosting service
@ LlamaIndex for orchestrating the entire RAG pipeline
@ Streamlit for interactive UI
Detailed Explanation on the Demo Functionality :
Crafted a Dockerfile to simplify the process and ensure easy reproducibility.
Set up models with Xinference and expose two ports for accessing them.
Leverage Streamlit for seamless file uploads and interactive communication with the chat engine.
5x faster doc embedding than OpenAI’s API.
Leveraging the power of GGML to offload models to the GPU, ensuring swift acceleration. Less long waits for returns.
- Source Code :