Real-time inference for Stable Diffusion - 0.88s latency. Covers AITemplate, nvFuser, TensorRT, FlashAttention. Join our Discord communty: https://discord.com/invite/TgHXuSJEk6
docker
notebook
cuda
inference
pytorch
automl
tensorrt
onnx
onnxruntime
stable-diffusion
aitemplate
nvfuser
-
Updated
Dec 4, 2023 - Jupyter Notebook