Skip to content
@NVIDIA

NVIDIA Corporation

Pinned Loading

  1. cuopt cuopt Public

    GPU accelerated decision optimization

    Cuda 806 156

  2. cuopt-examples cuopt-examples Public

    NVIDIA cuOpt examples for decision optimization

    Jupyter Notebook 431 73

  3. open-gpu-kernel-modules open-gpu-kernel-modules Public

    NVIDIA Linux open GPU kernel module source

    C 16.9k 1.7k

  4. aistore aistore Public

    AIStore: scalable storage for AI applications

    Go 1.8k 244

  5. nvidia-container-toolkit nvidia-container-toolkit Public

    Build and run containers leveraging NVIDIA GPUs

    Go 4.2k 506

  6. GenerativeAIExamples GenerativeAIExamples Public

    Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.

    Jupyter Notebook 3.9k 1k

Repositories

Showing 10 of 709 repositories
  • ncx-infra-controller-core Public

    NCX Infra Controller - Hardware Lifecycle Management and multitenant networking

    NVIDIA/ncx-infra-controller-core’s past year of commit activity
    Rust 114 Apache-2.0 74 128 (5 issues need help) 53 Updated Apr 8, 2026
  • aicr Public

    Tooling for optimized, validated, and reproducible GPU-accelerated AI runtime in Kubernetes

    NVIDIA/aicr’s past year of commit activity
    Go 258 Apache-2.0 27 18 10 Updated Apr 8, 2026
  • NVFlare Public

    NVIDIA Federated Learning Application Runtime Environment

    NVIDIA/NVFlare’s past year of commit activity
    Python 918 Apache-2.0 247 13 21 Updated Apr 8, 2026
  • Model-Optimizer Public

    A unified library of SOTA model optimization techniques like quantization, pruning, distillation, speculative decoding, etc. It compresses deep learning models for downstream deployment frameworks like TensorRT-LLM, TensorRT, vLLM, etc. to optimize inference speed.

    NVIDIA/Model-Optimizer’s past year of commit activity
    Python 2,397 Apache-2.0 336 60 124 Updated Apr 8, 2026
  • TensorRT-LLM Public

    TensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and supports state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in a performant way.

    NVIDIA/TensorRT-LLM’s past year of commit activity
    Python 13,316 2,261 569 667 Updated Apr 8, 2026
  • cuda-python Public

    CUDA Python: Performance meets Productivity

    NVIDIA/cuda-python’s past year of commit activity
    Cython 3,212 266 197 26 Updated Apr 8, 2026
  • k8s-device-plugin Public

    NVIDIA device plugin for Kubernetes

    NVIDIA/k8s-device-plugin’s past year of commit activity
    Go 3,717 Apache-2.0 803 57 58 Updated Apr 8, 2026
  • physicsnemo Public

    Open-source deep-learning framework for building, training, and fine-tuning deep learning models using state-of-the-art Physics-ML methods

    NVIDIA/physicsnemo’s past year of commit activity
    Python 2,644 Apache-2.0 634 21 40 Updated Apr 8, 2026
  • Megatron-LM Public

    Ongoing research training transformer models at scale

    NVIDIA/Megatron-LM’s past year of commit activity
    Python 15,951 3,790 343 (1 issue needs help) 349 Updated Apr 8, 2026
  • aistore Public

    AIStore: scalable storage for AI applications

    NVIDIA/aistore’s past year of commit activity
    Go 1,810 MIT 244 1 1 Updated Apr 8, 2026

Top languages

Loading…

Most used topics

Loading…