Skip to content
@NVIDIA

NVIDIA Corporation

Pinned Loading

  1. cuopt cuopt Public

    GPU accelerated decision optimization

    Cuda 448 77

  2. cuopt-examples cuopt-examples Public

    NVIDIA cuOpt examples for decision optimization

    Jupyter Notebook 367 51

  3. open-gpu-kernel-modules open-gpu-kernel-modules Public

    NVIDIA Linux open GPU kernel module source

    C 16.2k 1.5k

  4. aistore aistore Public

    AIStore: scalable storage for AI applications

    Go 1.6k 218

  5. nvidia-container-toolkit nvidia-container-toolkit Public

    Build and run containers leveraging NVIDIA GPUs

    Go 3.7k 416

  6. GenerativeAIExamples GenerativeAIExamples Public

    Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.

    Jupyter Notebook 3.4k 854

Repositories

Showing 10 of 607 repositories
  • KAI-Scheduler Public

    KAI Scheduler is an open source Kubernetes Native scheduler for AI workloads at large scale

    NVIDIA/KAI-Scheduler’s past year of commit activity
    Go 838 Apache-2.0 93 28 23 Updated Oct 5, 2025
  • warp Public

    A Python framework for accelerated simulation, data generation and spatial computing.

    NVIDIA/warp’s past year of commit activity
    Python 5,595 Apache-2.0 366 250 9 Updated Oct 5, 2025
  • NVFlare Public

    NVIDIA Federated Learning Application Runtime Environment

    NVIDIA/NVFlare’s past year of commit activity
    Python 801 Apache-2.0 211 13 18 Updated Oct 4, 2025
  • Fuser Public

    A Fusion Code Generator for NVIDIA GPUs (commonly known as "nvFuser")

    NVIDIA/Fuser’s past year of commit activity
    C++ 356 69 191 (15 issues need help) 189 Updated Oct 5, 2025
  • bionemo-framework Public

    BioNeMo Framework: For building and adapting AI models in drug discovery at scale

    NVIDIA/bionemo-framework’s past year of commit activity
    Jupyter Notebook 530 86 56 (1 issue needs help) 85 Updated Oct 5, 2025
  • TensorRT-LLM Public

    TensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and support state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in performant way.

    NVIDIA/TensorRT-LLM’s past year of commit activity
    C++ 11,772 Apache-2.0 1,779 721 372 Updated Oct 5, 2025
  • cuopt Public

    GPU accelerated decision optimization

    NVIDIA/cuopt’s past year of commit activity
    Cuda 448 Apache-2.0 77 87 10 Updated Oct 5, 2025
  • Megatron-LM Public

    Ongoing research training transformer models at scale

    NVIDIA/Megatron-LM’s past year of commit activity
    Python 13,747 3,138 308 119 Updated Oct 5, 2025
  • NeMo-Skills Public

    A project to improve skills of large language models

    NVIDIA/NeMo-Skills’s past year of commit activity
    Python 571 Apache-2.0 102 54 17 Updated Oct 5, 2025
  • TensorRT-Model-Optimizer Public

    A unified library of state-of-the-art model optimization techniques like quantization, pruning, distillation, speculative decoding, etc. It compresses deep learning models for downstream deployment frameworks like TensorRT-LLM or TensorRT to optimize inference speed.

    NVIDIA/TensorRT-Model-Optimizer’s past year of commit activity
    Python 1,423 Apache-2.0 168 119 (1 issue needs help) 23 Updated Oct 5, 2025