OneDiff: An out-of-the-box acceleration library for diffusion models.
-
Updated
May 22, 2024 - Python
CUDA® is a parallel computing platform and programming model developed by NVIDIA for general computing on graphical processing units (GPUs). With CUDA, developers are able to dramatically speed up computing applications by harnessing the power of GPUs.
OneDiff: An out-of-the-box acceleration library for diffusion models.
A Differentiable THB-spline module implemented in JAX and PyTorch
Containers for machine learning
ppl.cv is a high-performance image processing library of openPPL supporting x86 and cuda platforms.
Modular ZK(Zero Knowledge) backend accelerated by GPU
Open Voice OS Status Page
The open-source serverless GPU container runtime.
A high-throughput and memory-efficient inference and serving engine for LLMs
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
Vulkan render
A retargetable MLIR-based machine learning compiler and runtime toolkit.
Recreating PyTorch from scratch (C/C++, CUDA and Python, with GPU support and automatic differentiation!)
Pytorch domain library for recommendation systems
PyTorch/TorchScript/FX compiler for NVIDIA GPUs using TensorRT
A fast, scalable, high performance Gradient Boosting on Decision Trees library, used for ranking, classification, regression and other machine learning tasks for Python, R, Java, C++. Supports computation on CPU and GPU.
RAFT contains fundamental widely-used algorithms and primitives for machine learning and information retrieval. The algorithms are CUDA-accelerated and form building blocks for more easily writing high performance applications.
Created by Nvidia
Released June 23, 2007