Your Personalised Ai powered PDF buddy !!
Django-based RAG app to upload PDFs, index them into Pinecone using embeddings from Hugging Face and use ollama distilled model for RAG inference, and answer questions with Ollama zephyr:7b-alpha.
- Upload PDF and auto-split into chunks
- Embeddings via all-mpnet-base-v2
- Pinecone v3 serverless vector store
- RAG inference via local Ollama zephyr-7b-alpha (distilled model)
- Evidence contexts
- Python 3.10+ and PyTorch.
- Install Ollama and pull model:
ollama pull zephyr:7b-alpha
- Create and fill
.env(see.env.example). - Install dependencies:
pip install -r requirements.txt
- Django init:
python manage.py migratepython manage.py runserver
- Open http://127.0.0.1:8000
See .env.example for all settings:
PINECONE_API_KEYrequired.- If the Hugging Face model is gated, set
HUGGINGFACE_TOKEN. - Ensure Ollama is running locally at
OLLAMA_BASE_URL.