You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
It looks like on this line, you check if the custom layer norm op is installed. if so, this param is set to true. Following the call stack, that sets this param in the Flash-Attention package. That implementation here has moved to a Triton implementation.
However, later in the original hyena-DNA code, we are using the non-Triton function. Does that need to be swapped out?
In the Flash-attention repo here, there is now a note that the fused CUDA op has been replaced with a Triton op.
in light of that, is it now reasonable to remove from the dependencies section of this readme the suggestion to
pip install
the layer norm op?The text was updated successfully, but these errors were encountered: