I see this constraint of Flash Attention not being supported currently with Llama-3.2-11B-Vision-Instruct model. Any assistance here
· Sign up or log in to comment