Skip to content

update to use local LLM and embedding#83

Open
jcfortune wants to merge 19 commits intoNVIDIA:mainfrom
jcfortune:main
Open

update to use local LLM and embedding#83
jcfortune wants to merge 19 commits intoNVIDIA:mainfrom
jcfortune:main

Conversation

@jcfortune
Copy link

update to use local LLM and embedding - all changes from Gemini -
Summary of Final Architecture
By merging these changes, your Agri-Culture intelligence system now follows a professional microservices pattern:

AI Workbench Container: Runs the Python/Streamlit frontend and orchestrates the RAG logic.

LLM Sidecar: Your local Llama 3.1 8B running in "Eager Mode" on the Blackwell GPU.

Embedding Sidecar: Your native ARM64 300M model providing sub-millisecond vectorization.

- Updated Requested GPUs to '2'
- Modified files: compose.yaml
- Added pip packages: pydantic==2.10.6
- Updated Base Environment to 'nvidia/ai-workbench/python-cuda122:1.0.8'
- Modified files: compose.yaml, requirements.txt
- Added pip packages: pydantic==2.10.6
- Removed pip packages: pydantic==2.10.6
- Modified files: requirements.txt
- Updated Requested GPUs to '1'
- Updated Requested GPUs to '0'
- Modified files: code/upload-pdfs.ipynb, compose.yaml
- Updated Requested GPUs to '1'
- Modified files: .project/spec.yaml
updated for local LLM & embedding
change to local LLM & embedding
change to local LLM and embedding
updo changes...
updo changes
new update add LLM & embedding
new update to use local LLM and embedding
more changes
update to use local LMM & embedding
update to use local LLM & embedding
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant