This repository will soon be deleted as it's now deprecated. Please use kernels-community/layer-norm.

This CUDA extension implements fused dropout + residual + LayerNorm from the flash-attention repo.

Downloads last month
53
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support