Flash Attention Support

#41
by rameshch - opened

I see this constraint of Flash Attention not being supported currently with Llama-3.2-11B-Vision-Instruct model. Any assistance here

Sign up or log in to comment