FROM python:3.10.9 # Set the working directory to / WORKDIR / # Copy the current directory contents into the container at . COPY ./requirements.txt /requirements.txt RUN pip install transformers # Install pytorch starts RUN pip install torch #Install pytorch ends RUN pip install accelerate RUN pip install sentencepiece RUN pip install protobuf RUN pip install lxml # Install hugging face hub to download llama2 model #RUN pip install --upgrade huggingface_hub #RUN CMAKE_ARGS="-DLLAMA_BLAS=ON -DLLAMA_BLAS_VENDOR=OpenBLAS" pip install 'llama-cpp-python[server]' --upgrade --force-reinstall --no-cache-dir # Install requirements.txt RUN pip install --no-cache-dir --upgrade -r /requirements.txt # write access RUN useradd -m -u 1000 user USER user ENV HOME=/home/user \ PATH=/home/user/.local/bin:$PATH WORKDIR $HOME/app #RUN huggingface-cli download TheBloke/Llama-2-7b-Chat-GGUF llama-2-7b-chat.Q2_K.gguf --local-dir . --local-dir-use-symlinks False COPY --chown=user . $HOME/app # Start the FastAPI app on port 7860, the default port expected by Spaces CMD ["uvicorn", "question_paper:app", "--host", "0.0.0.0", "--port", "7860"]