- Python 3.10 or higher
- Pinecone as Vector Store
- Huggingface and Transformers for locally deploying LLM (We are using Google's Flan T5 Base model with 248M parameters (3GB) - so that my laptop doesn't die)
- Langchain for Conversation chain wrapper around the LLM
- Streamlit for the UI
- AWS Sagemaker for scaled up version
Create a virtual environment :
virtualenv .venv
Install requirements :
pip3 install requirements.txt
Keep your documents under doc_chat/assets
Indexing the documents to create embeddings and store it in Pinecone :
python3 indexing.py
Chat with the document :
streamlit run main.py
Sample user Q&A :
Fetching query-relevant context from documents :