This repository will soon be deleted as it's now deprecated. Please use kernels-community/layer-norm.
This CUDA extension implements fused dropout + residual + LayerNorm from the flash-attention repo.
- Downloads last month
- 53
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support