Skip to content
@NVIDIA

NVIDIA Corporation

Pinned Loading

  1. cuopt cuopt Public

    GPU accelerated decision optimization

    Cuda 723 129

  2. cuopt-examples cuopt-examples Public

    NVIDIA cuOpt examples for decision optimization

    Jupyter Notebook 414 66

  3. open-gpu-kernel-modules open-gpu-kernel-modules Public

    NVIDIA Linux open GPU kernel module source

    C 16.7k 1.6k

  4. aistore aistore Public

    AIStore: scalable storage for AI applications

    Go 1.8k 237

  5. nvidia-container-toolkit nvidia-container-toolkit Public

    Build and run containers leveraging NVIDIA GPUs

    Go 4.1k 482

  6. GenerativeAIExamples GenerativeAIExamples Public

    Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.

    Jupyter Notebook 3.8k 986

Repositories

Showing 10 of 679 repositories
  • Megatron-LM Public

    Ongoing research training transformer models at scale

    NVIDIA/Megatron-LM’s past year of commit activity
    Python 15,469 3,642 301 (1 issue needs help) 319 Updated Mar 2, 2026
  • TensorRT-LLM Public

    TensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and supports state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in a performant way.

    NVIDIA/TensorRT-LLM’s past year of commit activity
    Python 12,980 2,135 536 549 Updated Mar 2, 2026
  • OSMO Public

    The developer-first platform for scaling complex Physical AI workloads across heterogeneous compute—unifying training GPUs, simulation clusters, and edge devices in a simple YAML

    NVIDIA/OSMO’s past year of commit activity
    TypeScript 102 Apache-2.0 19 60 9 Updated Mar 2, 2026
  • TransformerEngine Public

    A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit and 4-bit floating point (FP8 and FP4) precision on Hopper, Ada and Blackwell GPUs, to provide better performance with lower memory utilization in both training and inference.

    NVIDIA/TransformerEngine’s past year of commit activity
    Python 3,177 Apache-2.0 651 239 127 Updated Mar 2, 2026
  • gpu-driver-container Public

    The NVIDIA GPU driver container allows the provisioning of the NVIDIA driver through the use of containers.

    NVIDIA/gpu-driver-container’s past year of commit activity
    Shell 160 Apache-2.0 77 23 36 Updated Mar 2, 2026
  • NVSentinel Public

    NVSentinel is a cross-platform fault remediation service designed to rapidly remediate runtime node-level issues in GPU-accelerated computing environments

    NVIDIA/NVSentinel’s past year of commit activity
    Go 192 Apache-2.0 49 46 21 Updated Mar 1, 2026
  • cuda-quantum Public

    C++ and Python support for the CUDA Quantum programming model for heterogeneous quantum-classical workflows

    NVIDIA/cuda-quantum’s past year of commit activity
    C++ 944 341 429 (16 issues need help) 106 Updated Mar 2, 2026
  • IsaacTeleop Public

    The unified framework for sim & real robot teleoperation

    NVIDIA/IsaacTeleop’s past year of commit activity
    Python 7 0 1 10 Updated Mar 2, 2026
  • cudaqx Public

    Accelerated libraries for quantum-classical computing built on CUDA-Q.

    NVIDIA/cudaqx’s past year of commit activity
    C++ 85 53 28 (1 issue needs help) 18 Updated Mar 2, 2026
  • Model-Optimizer Public

    A unified library of SOTA model optimization techniques like quantization, pruning, distillation, speculative decoding, etc. It compresses deep learning models for downstream deployment frameworks like TensorRT-LLM, TensorRT, vLLM, etc. to optimize inference speed.

    NVIDIA/Model-Optimizer’s past year of commit activity
    Python 2,078 Apache-2.0 284 68 97 Updated Mar 2, 2026