Skip to content
@NVIDIA

NVIDIA Corporation

Pinned Loading

  1. cuopt cuopt Public

    NVIDIA cuOpt is an open-source GPU-accelerated optimization engine delivering near real-time solutions for complex decision-making challenges.

    Cuda 186 24

  2. cuopt-examples cuopt-examples Public

    NVIDIA cuOpt examples for decision optimization

    Jupyter Notebook 326 45

  3. open-gpu-kernel-modules open-gpu-kernel-modules Public

    NVIDIA Linux open GPU kernel module source

    C 15.9k 1.4k

  4. aistore aistore Public

    AIStore: scalable storage for AI applications

    Go 1.5k 211

  5. nvidia-container-toolkit nvidia-container-toolkit Public

    Build and run containers leveraging NVIDIA GPUs

    Go 3.3k 355

  6. GenerativeAIExamples GenerativeAIExamples Public

    Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.

    Jupyter Notebook 3.2k 766

Repositories

Showing 10 of 580 repositories
  • Megatron-LM Public

    Ongoing research training transformer models at scale

    NVIDIA/Megatron-LM’s past year of commit activity
    Python 12,600 2,853 308 201 Updated Jun 18, 2025
  • kvpress Public

    LLM KV cache compression made easy

    NVIDIA/kvpress’s past year of commit activity
    Python 509 Apache-2.0 40 3 0 Updated Jun 18, 2025
  • cccl Public

    CUDA Core Compute Libraries

    NVIDIA/cccl’s past year of commit activity
    C++ 1,692 225 981 (5 issues need help) 123 Updated Jun 18, 2025
  • nvidia-container-toolkit Public

    Build and run containers leveraging NVIDIA GPUs

    NVIDIA/nvidia-container-toolkit’s past year of commit activity
    Go 3,324 Apache-2.0 355 381 30 Updated Jun 18, 2025
  • KAI-Scheduler Public

    KAI Scheduler is an open source Kubernetes Native scheduler for AI workloads at large scale

    NVIDIA/KAI-Scheduler’s past year of commit activity
    Go 649 Apache-2.0 70 18 (2 issues need help) 14 Updated Jun 18, 2025
  • TensorRT-LLM Public

    TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and support state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in performant way.

    NVIDIA/TensorRT-LLM’s past year of commit activity
    C++ 10,769 Apache-2.0 1,504 616 274 Updated Jun 18, 2025
  • cloudai Public

    CloudAI Benchmark Framework

    NVIDIA/cloudai’s past year of commit activity
    Python 66 Apache-2.0 29 1 16 Updated Jun 18, 2025
  • torch-harmonics Public

    Differentiable signal processing on the sphere for PyTorch

    NVIDIA/torch-harmonics’s past year of commit activity
    Jupyter Notebook 469 44 4 1 Updated Jun 18, 2025
  • JAX-Toolbox Public

    JAX-Toolbox

    NVIDIA/JAX-Toolbox’s past year of commit activity
    Python 311 Apache-2.0 60 79 39 Updated Jun 18, 2025
  • grove Public

    Kubernetes enhancements for Network Topology Aware Gang Scheduling & Autoscaling

    NVIDIA/grove’s past year of commit activity
    Go 6 Apache-2.0 3 11 1 Updated Jun 18, 2025