[Cache Request] meta-llama/Llama-2-7b-chat-hf
#51
by
naveen1601datalyticsfoundry
- opened
Please add the following model to the neuron cache
Several cached configurations are actually available fro LLama.
https://huggingface.co/aws-neuron/optimum-neuron-cache/blob/main/inference-cache-config/llama.json
dacorvo
changed discussion status to
closed