shreyajn commited on
Commit
e706b0f
1 Parent(s): bb780dd

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +5 -3
README.md CHANGED
@@ -15,11 +15,13 @@ tags:
15
  # Llama-v3.2-3B-Chat: Optimized for Mobile Deployment
16
  ## State-of-the-art large language model useful on a variety of language understanding and generation tasks
17
 
 
18
  Llama 3 is a family of LLMs. The "Chat" at the end indicates that the model is optimized for chatbot-like dialogue. The model is quantized to w4a16 (4-bit weights and 16-bit activations) and part of the model is quantized to w8a16 (8-bit weights and 16-bit activations) making it suitable for on-device deployment. For Prompt and output length specified below, the time to first token is Llama-PromptProcessor-Quantized's latency and average time per addition token is Llama-TokenGenerator-Quantized's latency.
19
 
20
- This is based on the implementation of Llama-v3.2-3B-Chat found
21
- [here]({source_repo}). More details on model performance
22
- accross various devices, can be found [here](https://aihub.qualcomm.com/models/llama_v3_2_3b_chat_quantized).
 
23
 
24
  ### Model Details
25
 
 
15
  # Llama-v3.2-3B-Chat: Optimized for Mobile Deployment
16
  ## State-of-the-art large language model useful on a variety of language understanding and generation tasks
17
 
18
+
19
  Llama 3 is a family of LLMs. The "Chat" at the end indicates that the model is optimized for chatbot-like dialogue. The model is quantized to w4a16 (4-bit weights and 16-bit activations) and part of the model is quantized to w8a16 (8-bit weights and 16-bit activations) making it suitable for on-device deployment. For Prompt and output length specified below, the time to first token is Llama-PromptProcessor-Quantized's latency and average time per addition token is Llama-TokenGenerator-Quantized's latency.
20
 
21
+ This model is an implementation of Posenet-Mobilenet found [here](https://github.com/meta-llama/llama3/tree/main).
22
+
23
+
24
+ More details on model performance accross various devices, can be found [here](https://aihub.qualcomm.com/models/llama_v3_2_3b_chat_quantized).
25
 
26
  ### Model Details
27