Skip to content
@NVIDIA

NVIDIA Corporation

Pinned Loading

  1. cuopt cuopt Public

    GPU accelerated decision optimization

    Cuda 539 89

  2. cuopt-examples cuopt-examples Public

    NVIDIA cuOpt examples for decision optimization

    Jupyter Notebook 377 55

  3. open-gpu-kernel-modules open-gpu-kernel-modules Public

    NVIDIA Linux open GPU kernel module source

    C 16.3k 1.5k

  4. aistore aistore Public

    AIStore: scalable storage for AI applications

    Go 1.6k 222

  5. nvidia-container-toolkit nvidia-container-toolkit Public

    Build and run containers leveraging NVIDIA GPUs

    Go 3.8k 430

  6. GenerativeAIExamples GenerativeAIExamples Public

    Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.

    Jupyter Notebook 3.6k 901

Repositories

Showing 10 of 625 repositories
  • Fuser Public

    A Fusion Code Generator for NVIDIA GPUs (commonly known as "nvFuser")

    NVIDIA/Fuser’s past year of commit activity
    C++ 360 69 204 (15 issues need help) 194 Updated Nov 9, 2025
  • NVSentinel Public

    NVSentinel is a cross-platform fault remediation service designed to rapidly remediate runtime node-level issues in GPU-accelerated computing environments

    NVIDIA/NVSentinel’s past year of commit activity
    Go 68 Apache-2.0 14 21 7 Updated Nov 9, 2025
  • cccl Public

    CUDA Core Compute Libraries

    NVIDIA/cccl’s past year of commit activity
  • k8s-nim-operator Public

    An Operator for deployment and maintenance of NVIDIA NIMs and NeMo microservices in a Kubernetes environment.

    NVIDIA/k8s-nim-operator’s past year of commit activity
    Go 133 Apache-2.0 34 6 29 Updated Nov 9, 2025
  • JAX-Toolbox Public

    JAX-Toolbox

    NVIDIA/JAX-Toolbox’s past year of commit activity
    Python 359 Apache-2.0 66 80 34 Updated Nov 9, 2025
  • topograph Public

    A toolkit for discovering cluster network topology.

    NVIDIA/topograph’s past year of commit activity
    Go 79 Apache-2.0 6 2 1 Updated Nov 9, 2025
  • TensorRT-LLM Public

    TensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and supports state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in a performant way.

    NVIDIA/TensorRT-LLM’s past year of commit activity
    C++ 12,078 Apache-2.0 1,850 744 421 Updated Nov 9, 2025
  • TensorRT-Model-Optimizer Public

    A unified library of state-of-the-art model optimization techniques like quantization, pruning, distillation, speculative decoding, etc. It compresses deep learning models for downstream deployment frameworks like TensorRT-LLM or TensorRT to optimize inference speed.

    NVIDIA/TensorRT-Model-Optimizer’s past year of commit activity
    Python 1,523 Apache-2.0 192 58 46 Updated Nov 9, 2025
  • doca-sosreport Public Forked from sosreport/sos

    A unified tool for collecting system logs and other debug information

    NVIDIA/doca-sosreport’s past year of commit activity
    Python 4 GPL-2.0 595 0 2 Updated Nov 9, 2025
  • k8s-test-infra Public

    K8s-test-infra

    NVIDIA/k8s-test-infra’s past year of commit activity
    Go 10 Apache-2.0 8 0 5 Updated Nov 9, 2025