Cuda kernels created with kernels library
attention-int8 : A high-performance INT8 fused attention kernel for diffusion transformers.
Cuda kernels created with kernels library
attention-int8 : A high-performance INT8 fused attention kernel for diffusion transformers.