|
127 | 127 | "RAG_BASE_URL = f\"http://{IPADDRESS}:{RAG_SERVER_PORT}\"\n", |
128 | 128 | "INGESTOR_BASE_URL = f\"http://{IPADDRESS}:{INGESTOR_SERVER_PORT}\"\n", |
129 | 129 | "\n", |
| 130 | + "# NIM services to deploy (excludes nim-llm and vlm-ms since we use NVIDIA-hosted endpoints)\n", |
| 131 | + "NIM_SERVICES = (\n", |
| 132 | + " \"nemoretriever-embedding-ms \"\n", |
| 133 | + " \"nemoretriever-ranking-ms \"\n", |
| 134 | + " \"page-elements \"\n", |
| 135 | + " \"graphic-elements \"\n", |
| 136 | + " \"table-structure \"\n", |
| 137 | + " \"nemoretriever-ocr\"\n", |
| 138 | + ")\n", |
| 139 | + "\n", |
130 | 140 | "\n", |
131 | 141 | "# =============================================================================\n", |
132 | 142 | "# DOCKER COMPOSE HELPERS\n", |
|
520 | 530 | " print(\"DEPLOYING NVIDIA RAG BLUEPRINT\")\n", |
521 | 531 | " print(\"=\" * 60)\n", |
522 | 532 | " \n", |
523 | | - " print(\"\\n[1/4] NIM Microservices...\")\n", |
524 | | - " docker_compose(\"deploy/compose/nims.yaml\", \"pull\", \"-q\")\n", |
525 | | - " docker_compose(\"deploy/compose/nims.yaml\", \"up\", \"-d\")\n", |
| 533 | + " print(\"\\n[1/4] NIM Microservices (no local LLM/VLM)...\")\n", |
| 534 | + " docker_compose(\"deploy/compose/nims.yaml\", \"pull\", f\"-q {NIM_SERVICES}\")\n", |
| 535 | + " docker_compose(\"deploy/compose/nims.yaml\", \"up\", f\"-d {NIM_SERVICES}\")\n", |
526 | 536 | " print(\"-\" * 60)\n", |
527 | 537 | " print(\"\\n[2/4] Vector Database...\")\n", |
528 | 538 | " docker_compose(\"deploy/compose/vectordb.yaml\", \"pull\", \"-q\")\n", |
|
667 | 677 | " print(\" ❌ FAIL: nvidia-smi not found\")\n", |
668 | 678 | "\n", |
669 | 679 | "# ─────────────────────────────────────────────────────────────────────────────\n", |
670 | | - "# [3] NVIDIA Driver Version (need 560+)\n", |
| 680 | + "# [3] NVIDIA Driver Version (need 580+)\n", |
671 | 681 | "# ─────────────────────────────────────────────────────────────────────────────\n", |
672 | | - "print(\"\\n[3] NVIDIA Driver Version (need 560+):\")\n", |
| 682 | + "print(\"\\n[3] NVIDIA Driver Version (need 580+):\")\n", |
673 | 683 | "try:\n", |
674 | 684 | " result = subprocess.run([\"nvidia-smi\", \"-q\"], capture_output=True, text=True)\n", |
675 | 685 | " if result.returncode == 0:\n", |
676 | 686 | " match = re.search(r\"Driver Version\\s*:\\s*(\\d+)\", result.stdout)\n", |
677 | 687 | " if match:\n", |
678 | 688 | " driver_version = int(match.group(1))\n", |
679 | 689 | " print(f\" Driver Version: {driver_version}\")\n", |
680 | | - " if driver_version >= 560:\n", |
| 690 | + " if driver_version >= 580:\n", |
681 | 691 | " print(\" ✅ PASS\")\n", |
682 | 692 | " else:\n", |
683 | | - " errors.append(f\"Driver version {driver_version} < 560 required\")\n", |
684 | | - " print(f\" ❌ FAIL: Version {driver_version} < 560\")\n", |
| 693 | + " errors.append(f\"Driver version {driver_version} < 580 required\")\n", |
| 694 | + " print(f\" ❌ FAIL: Version {driver_version} < 580\")\n", |
685 | 695 | " else:\n", |
686 | 696 | " print(\" Unable to parse driver version\")\n", |
687 | 697 | "except FileNotFoundError:\n", |
688 | 698 | " print(\" ❌ FAIL: nvidia-smi not found\")\n", |
689 | 699 | "\n", |
690 | 700 | "# ─────────────────────────────────────────────────────────────────────────────\n", |
691 | | - "# [4] CUDA Version (need 12.4+)\n", |
| 701 | + "# [4] CUDA Version (need 13+)\n", |
692 | 702 | "# ─────────────────────────────────────────────────────────────────────────────\n", |
693 | | - "print(\"\\n[4] CUDA Version (need 12.4+):\")\n", |
| 703 | + "print(\"\\n[4] CUDA Version (need 13+):\")\n", |
694 | 704 | "try:\n", |
695 | 705 | " result = subprocess.run([\"nvidia-smi\", \"-q\"], capture_output=True, text=True)\n", |
696 | 706 | " if result.returncode == 0:\n", |
697 | 707 | " match = re.search(r\"CUDA Version\\s*:\\s*(\\d+\\.\\d+)\", result.stdout)\n", |
698 | 708 | " if match:\n", |
699 | 709 | " cuda_version = float(match.group(1))\n", |
700 | 710 | " print(f\" CUDA Version: {cuda_version}\")\n", |
701 | | - " if cuda_version >= 12.4:\n", |
| 711 | + " if cuda_version >= 13:\n", |
702 | 712 | " print(\" ✅ PASS\")\n", |
703 | 713 | " else:\n", |
704 | | - " errors.append(f\"CUDA version {cuda_version} < 12.4 required\")\n", |
705 | | - " print(f\" ❌ FAIL: Version {cuda_version} < 12.4\")\n", |
| 714 | + " errors.append(f\"CUDA version {cuda_version} < 13 required\")\n", |
| 715 | + " print(f\" ❌ FAIL: Version {cuda_version} < 13\")\n", |
706 | 716 | " else:\n", |
707 | 717 | " print(\" Unable to parse CUDA version\")\n", |
708 | 718 | "except FileNotFoundError:\n", |
|
1525 | 1535 | "metadata": {}, |
1526 | 1536 | "source": [ |
1527 | 1537 | "### 4.5 Study Extracted Results\n", |
| 1538 | + "⚠️ **Only run this cell if you ran cell 2.3a Save Extracted Content**\n", |
1528 | 1539 | "\n", |
1529 | 1540 | "When `APP_NVINGEST_SAVETODISK=True` is enabled, the ingestion pipeline saves the extracted results to:\n", |
1530 | 1541 | "```\n", |
|
0 commit comments