textgeneration / Dockerfile
Yash Sachdeva
llm download
2d5d217
raw
history blame
910 Bytes
FROM python:3.10.9
# Set the working directory to /
WORKDIR /
# Copy the current directory contents into the container at .
COPY . .
RUN pip install transformers
# Install pytorch starts
RUN pip install torch
#Install pytorch ends
RUN pip install accelerate
# Install hugging face hub to download llama2 model
RUN pip install --upgrade huggingface_hub
RUN CMAKE_ARGS="-DLLAMA_BLAS=ON -DLLAMA_BLAS_VENDOR=OpenBLAS" pip install 'llama-cpp-python[server]' --upgrade --force-reinstall --no-cache-dir
# Install requirements.txt
RUN pip install --no-cache-dir --upgrade -r /requirements.txt
# write access
RUN useradd -m -u 1000 user
USER user
ENV HOME=/home/user \
PATH=/home/user/.local/bin:$PATH
WORKDIR $HOME/app
COPY --chown=user . $HOME/app
# Start the FastAPI app on port 7860, the default port expected by Spaces
CMD ["uvicorn", "question_paper:app", "--host", "0.0.0.0", "--port", "7860"]