Spaces:
Sleeping
Sleeping
FROM pytorch/pytorch:2.0.1-cuda11.7-cudnn8-runtime | |
ENV DEBIAN_FRONTEND=noninteractive | |
# Set the MKL_THREADING_LAYER environment variable to GNU | |
ENV MKL_THREADING_LAYER=GNU | |
# Install Git, OpenGL libraries, and libglib2.0 | |
RUN apt-get update && apt-get install -y git libgl1-mesa-glx libglib2.0-0 | |
RUN apt-get update && apt-get install -y ninja-build | |
# Install necessary dependencies, including CMake, a C++ compiler, and others | |
RUN apt-get update && apt-get install -y unzip ffmpeg cmake g++ build-essential espeak-ng aria2 | |
# Set up a new user named "user" with user ID 1000 | |
RUN useradd -m -u 1000 user | |
# Switch to the "user" user | |
USER user | |
# Set environment variables | |
ENV HOME=/home/user \ | |
CUDA_HOME=/usr/local/cuda \ | |
PATH=/home/user/.local/bin:$PATH \ | |
LD_LIBRARY_PATH=${CUDA_HOME}/lib64:${LD_LIBRARY_PATH} \ | |
LIBRARY_PATH=${CUDA_HOME}/lib64/stubs:${LIBRARY_PATH} \ | |
PYTHONPATH=$HOME/app \ | |
PYTHONUNBUFFERED=1 \ | |
GRADIO_ALLOW_FLAGGING=never \ | |
GRADIO_NUM_PORTS=1 \ | |
GRADIO_SERVER_NAME=0.0.0.0 \ | |
GRADIO_THEME=huggingface \ | |
GRADIO_SHARE=False \ | |
SYSTEM=spaces | |
# Set the working directory to the user's home directory | |
WORKDIR $HOME/app | |
# Clone your repository or add your code to the container | |
RUN git clone -b main https://github.com/fffiloni/AniPortrait $HOME/app | |
# Install dependencies | |
#COPY requirements.txt $HOME/app/requirements.txt | |
RUN pip install --no-cache-dir -r requirements.txt | |
# Download checkpoint files using aria2 | |
RUN aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/camenduru/AniPortrait/resolve/main/audio2mesh.pt -d $HOME/app/pretrained_model -o audio2mesh.pt | |
RUN aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/camenduru/AniPortrait/resolve/main/denoising_unet.pth -d $HOME/app/pretrained_model -o denoising_unet.pth | |
RUN aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/camenduru/AniPortrait/resolve/main/motion_module.pth -d $HOME/app/pretrained_model -o motion_module.pth | |
RUN aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/camenduru/AniPortrait/resolve/main/pose_guider.pth -d $HOME/app/pretrained_model -o pose_guider.pth | |
RUN aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/camenduru/AniPortrait/resolve/main/reference_unet.pth -d $HOME/app/pretrained_model -o reference_unet.pth | |
RUN aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/camenduru/champ/resolve/main/champ/denoising_unet.pth -d $HOME/app/pretrained_model -o denoising_unet.pth | |
RUN aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/camenduru/champ/resolve/main/champ/guidance_encoder_depth.pth -d $HOME/app/pretrained_model -o guidance_encoder_depth.pth | |
RUN aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/camenduru/champ/resolve/main/champ/guidance_encoder_dwpose.pth -d $HOME/app/pretrained_model -o guidance_encoder_dwpose.pth | |
RUN aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/camenduru/champ/resolve/main/champ/guidance_encoder_normal.pth -d $HOME/app/pretrained_model -o guidance_encoder_normal.pth | |
RUN aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/camenduru/champ/resolve/main/champ/guidance_encoder_semantic_map.pth -d $HOME/app/pretrained_model -o guidance_encoder_semantic_map.pth | |
RUN aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/camenduru/champ/resolve/main/champ/motion_module.pth -d $HOME/app/pretrained_model -o motion_module.pth | |
RUN aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/camenduru/champ/resolve/main/champ/reference_unet.pth -d $HOME/app/pretrained_model -o reference_unet.pth | |
RUN aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/camenduru/champ/raw/main/image_encoder/config.json -d $HOME/app/pretrained_model/image_encoder -o config.json | |
RUN aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/camenduru/champ/resolve/main/image_encoder/pytorch_model.bin -d $HOME/app/pretrained_model/image_encoder -o pytorch_model.bin | |
RUN aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/camenduru/champ/raw/main/sd-vae-ft-mse/config.json -d $HOME/app/pretrained_model/sd-vae-ft-mse -o config.json | |
RUN aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/camenduru/champ/resolve/main/sd-vae-ft-mse/diffusion_pytorch_model.bin -d $HOME/app/pretrained_model/sd-vae-ft-mse -o diffusion_pytorch_model.bin | |
RUN aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/camenduru/champ/raw/main/stable-diffusion-v1-5/feature_extractor/preprocessor_config.json -d $HOME/app/pretrained_model/stable-diffusion-v1-5/feature_extractor -o preprocessor_config.json | |
RUN aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/camenduru/champ/raw/main/stable-diffusion-v1-5/model_index.json -d $HOME/app/pretrained_model/stable-diffusion-v1-5 -o model_index.json | |
RUN aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/camenduru/champ/raw/main/stable-diffusion-v1-5/unet/config.json -d $HOME/app/pretrained_model/stable-diffusion-v1-5/unet -o config.json | |
RUN aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/camenduru/champ/resolve/main/stable-diffusion-v1-5/unet/diffusion_pytorch_model.bin -d $HOME/app/pretrained_model/stable-diffusion-v1-5/unet -o diffusion_pytorch_model.bin | |
RUN aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/camenduru/champ/raw/main/stable-diffusion-v1-5/v1-inference.yaml -d $HOME/app/pretrained_model/stable-diffusion-v1-5 -o v1-inference.yaml | |
RUN aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/facebook/wav2vec2-base-960h/raw/main/config.json -d $HOME/app/pretrained_model/wav2vec2-base-960h -o config.json | |
RUN aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/facebook/wav2vec2-base-960h/raw/main/feature_extractor_config.json -d $HOME/app/pretrained_model/wav2vec2-base-960h -o feature_extractor_config.json | |
RUN aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/facebook/wav2vec2-base-960h/raw/main/preprocessor_config.json -d $HOME/app/pretrained_model/wav2vec2-base-960h -o preprocessor_config.json | |
RUN aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/facebook/wav2vec2-base-960h/resolve/main/pytorch_model.bin -d $HOME/app/pretrained_model/wav2vec2-base-960h -o pytorch_model.bin | |
RUN aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/facebook/wav2vec2-base-960h/raw/main/special_tokens_map.json -d $HOME/app/pretrained_model/wav2vec2-base-960h -o special_tokens_map.json | |
RUN aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/facebook/wav2vec2-base-960h/resolve/main/tf_model.h5 -d $HOME/app/pretrained_model/wav2vec2-base-960h -o tf_model.h5 | |
RUN aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/facebook/wav2vec2-base-960h/raw/main/tokenizer_config.json -d $HOME/app/pretrained_model/wav2vec2-base-960h -o tokenizer_config.json | |
RUN aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/facebook/wav2vec2-base-960h/raw/main/vocab.json -d $HOME/app/pretrained_model/wav2vec2-base-960h -o vocab.json | |
# Set the environment variable to specify the GPU device | |
ENV CUDA_DEVICE_ORDER=PCI_BUS_ID | |
ENV CUDA_VISIBLE_DEVICES=0 | |
RUN mkdir examples | |
COPY examples examples | |
COPY app_hf.py . | |
# Run your app.py script | |
CMD ["python", "app_hf.py"] |