seanpedrickcase commited on
Commit
08eb30d
1 Parent(s): 34f1e83

Updated Dockerfile with latest packages

Browse files
Files changed (2) hide show
  1. Dockerfile +4 -4
  2. run_cmd_line_example_command.txt +1 -1
Dockerfile CHANGED
@@ -16,10 +16,10 @@ WORKDIR /src
16
  COPY requirements_aws.txt .
17
 
18
  RUN pip install torch==2.4.0+cpu --target=/install --index-url https://download.pytorch.org/whl/cpu \
19
- && pip install --no-cache-dir --target=/install sentence-transformers==3.0.1 --no-deps \
20
- && pip install --no-cache-dir --target=/install bertopic==0.16.2 --no-deps \
21
  && pip install --no-cache-dir --target=/install -r requirements_aws.txt \
22
- && pip install --no-cache-dir --target=/install gradio==4.44.0
23
 
24
  # Add /install to the PYTHONPATH
25
  ENV PYTHONPATH="/install:${PYTHONPATH}"
@@ -44,7 +44,7 @@ RUN mkdir -p /home/user/app/output /home/user/.cache/huggingface/hub /home/user/
44
  && chown -R user:user /home/user
45
 
46
  # Download the quantised phi model directly with curl. Changed at it is so big - not loaded
47
- #RUN curl -L -o /home/user/app/model/rep/Phi-3.1-mini-128k-instruct-Q4_K_M.gguf https://huggingface.co/bartowski/Phi-3.1-mini-128k-instruct-GGUF/tree/main/Phi-3.1-mini-128k-instruct-Q4_K_M.gguf
48
 
49
  # Copy models from the builder stage
50
  COPY --from=builder /model/rep /home/user/app/model/rep
 
16
  COPY requirements_aws.txt .
17
 
18
  RUN pip install torch==2.4.0+cpu --target=/install --index-url https://download.pytorch.org/whl/cpu \
19
+ && pip install --no-cache-dir --target=/install sentence-transformers==3.2.0 --no-deps \
20
+ && pip install --no-cache-dir --target=/install bertopic==0.16.4 --no-deps \
21
  && pip install --no-cache-dir --target=/install -r requirements_aws.txt \
22
+ && pip install --no-cache-dir --target=/install gradio==5.6.0
23
 
24
  # Add /install to the PYTHONPATH
25
  ENV PYTHONPATH="/install:${PYTHONPATH}"
 
44
  && chown -R user:user /home/user
45
 
46
  # Download the quantised phi model directly with curl. Changed at it is so big - not loaded
47
+ #RUN curl -L -o /home/user/app/model/rep/Llama-3.2-3B-Instruct-Q5_K_M.gguf https://huggingface.co/bartowski/Llama-3.2-3B-Instruct-GGUF/tree/main/Llama-3.2-3B-Instruct-Q5_K_M.gguf
48
 
49
  # Copy models from the builder stage
50
  COPY --from=builder /model/rep /home/user/app/model/rep
run_cmd_line_example_command.txt CHANGED
@@ -1,3 +1,3 @@
1
  # To just get embeddings out
2
 
3
- python run_from_cmd_line.py --data_file "C:\Users\SPedrickCase\OneDrive - Lambeth Council\Apps\topic_modelling\examples\combined_case_notes.csv" --in_colnames "Case Note" --clean_text Yes --drop_duplicate_text No --anonymise_drop Yes --split_sentence_drop No --custom_regex_file "C:\Users\SPedrickCase\OneDrive - Lambeth Council\Apps\topic_modelling\examples\regex_to_remove.csv" --embeddings_high_quality_mode Yes --return_only_embeddings_drop Yes --output_folder "C:\Users\SPedrickCase\OneDrive - Lambeth Council\2024\ASC Project\asc_predict\apps\topic_modelling\output"
 
1
  # To just get embeddings out
2
 
3
+ python run_from_cmd_line.py --data_file "C:\folder\combined_case_notes.csv" --in_colnames "Case Note" --clean_text Yes --drop_duplicate_text No --anonymise_drop Yes --split_sentence_drop No --custom_regex_file "C:\folder\regex_to_remove.csv" --embeddings_high_quality_mode Yes --return_only_embeddings_drop Yes --output_folder "C:\folder\apps\topic_modelling\output"