[Cache Request] meta-llama/Meta-Llama-3-8B

#65
by huntingcarlisle - opened
AWS Inferentia and Trainium org

Please add the following model to the neuron cache

AWS Inferentia and Trainium org

This model is already cached. Please use optimum-cli neuron cache lookup meta-llama/Meta-Llama-3-8B to see the list of cached configurations.

dacorvo changed discussion status to closed

Sign up or log in to comment