adarshxs commited on
Commit
005705b
1 Parent(s): 136214c

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +57 -0
README.md ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ base_model: meta-llama/Meta-Llama-3.1-8B-Instruct
6
+ pipeline_tag: image-text-to-text
7
+ ---
8
+
9
+ # Llama 3.1 Vision by Capx AI
10
+
11
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/644bf6ef778ecbfb977e8e84/3D-oR8GazhHTaA-kVLNDk.png)
12
+
13
+ Read more on: https://huggingface.co/blog/adarshxs/capx-vision
14
+
15
+
16
+
17
+ ## Directions to Run Inference:
18
+ **Minimum requirements to run Inference is an A100 40GB GPU.**
19
+
20
+ - Clone our fork of the Bunny by BAAI repository here: https://github.com/adarshxs/Capx-Llama3.1-Vision
21
+ - Create a conda virtual environment
22
+ ```bash
23
+ conda create -n capx python=3.10
24
+ conda activate capx
25
+ ```
26
+ - Install the following
27
+ ```bash
28
+ pip install --upgrade pip # enable PEP 660 support
29
+ pip install transformers
30
+ pip install torch torchvision xformers --index-url https://download.pytorch.org/whl/cu118
31
+
32
+ # Installing APEX
33
+ pip install ninja
34
+ git clone https://github.com/NVIDIA/apex
35
+ cd apex
36
+ pip install -v --disable-pip-version-check --no-cache-dir --no-build-isolation --global-option="--cpp_ext" --global-option="--cuda_ext" ./
37
+ cd ..
38
+
39
+ # Installing Flash Attn
40
+ pip install packaging
41
+ pip install flash-attn --no-build-isolation
42
+
43
+ # Clone the inference Repo
44
+ git clone https://github.com/adarshxs/Capx-Llama3.1-Vision
45
+ cd Capx-Llama3.1-Vision
46
+ pip install -e .
47
+ ```
48
+ - Run cli server:
49
+ ```bash
50
+ python -m bunny.serve.cli \
51
+ --model-path Capx/Llama-3.1-Vision \
52
+ --model-type llama3.1-8b \
53
+ --image-file /path/to/image \
54
+ --conv-mode llama
55
+ ```
56
+
57
+ We thank the amazing team at BAAI, for their Bunny project, upon which this was built and Meta AI for their Llama 3.1 model!