You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
It would be wonderful if there is a support for this feature which is equivalent to flash attention v2.3.
This would also support Mistral-7B model as well, which is one of the best opensource 7B model architecture.
May I know is there a plan to bump flash attention v2.0.4 ROCm to v2.3?
tuyaao, jamestwhedbee, ekazakos, RichardFevrier and ehartford