textgeneration / Dockerfile
Yash Sachdeva
download llama
6db4d36
raw
history blame
1.04 kB
FROM python:3.10.9
# Set the working directory to /
WORKDIR /
# Copy the current directory contents into the container at .
COPY . .
RUN pip install transformers
# Install pytorch starts
RUN pip install torch
#Install pytorch ends
RUN pip install accelerate
# Install hugging face hub to download llama2 model
RUN pip install --upgrade huggingface_hub
RUN CMAKE_ARGS="-DLLAMA_BLAS=ON -DLLAMA_BLAS_VENDOR=OpenBLAS" pip install 'llama-cpp-python[server]' --upgrade --force-reinstall --no-cache-dir
# Install requirements.txt
RUN pip install --no-cache-dir --upgrade -r /requirements.txt
# write access
RUN useradd -m -u 1000 user
USER user
ENV HOME=/home/user \
PATH=/home/user/.local/bin:$PATH
WORKDIR $HOME/app
RUN huggingface-cli download TheBloke/Llama-2-7b-Chat-GGUF llama-2-7b-chat.Q4_K_M.gguf --local-dir . --local-dir-use-symlinks False
COPY --chown=user . $HOME/app
# Start the FastAPI app on port 7860, the default port expected by Spaces
CMD ["uvicorn", "question_paper:app", "--host", "0.0.0.0", "--port", "7860"]