|
--- |
|
base_model: appvoid/arco |
|
license: apache-2.0 |
|
tags: |
|
- llama-cpp |
|
- gguf-my-repo |
|
--- |
|
|
|
# TobDeBer/arco-Q4_K_M-GGUF |
|
This model was converted to Big Endian Q4_K_M GGUF format from [`appvoid/arco`](https://huggingface.co/appvoid/arco) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space. |
|
Refer to the [original model card](https://huggingface.co/appvoid/arco) for more details on the model. |
|
|
|
|
|
# Container Repository for CPU adaptations of Inference code |
|
|
|
## Variants for Inference |
|
|
|
### Slim container |
|
|
|
- run std binaries |
|
|
|
|
|
### CPUdiffusion |
|
|
|
- inference diffusion models on CPU |
|
- include CUDAonCPU stack |
|
|
|
### Diffusion container |
|
|
|
- run diffusion app.py variants |
|
- support CPU and CUDA |
|
- include Flux |
|
|
|
### Slim CUDA container |
|
|
|
- run CUDA binaries |
|
|
|
|
|
## Variants for Build |
|
|
|
### Llama.cpp build container |
|
|
|
- build llama-cli-static |
|
- build llama-server-static |
|
|
|
### sd build container |
|
|
|
- build sd |
|
- optional: build sd-server |
|
|
|
### CUDA build container |
|
|
|
- build cuda binaries |
|
- support sd_cuda |
|
|
|
|
|
|