AurelioAguirre commited on
Commit
15c704c
·
1 Parent(s): 11c65a0

Dockerfile set to Python3.12-mini

Browse files
Files changed (2) hide show
  1. Dockerfile +2 -1
  2. main/logs/llm_api.log +21 -0
Dockerfile CHANGED
@@ -1,4 +1,5 @@
1
- FROM python:3.10
 
2
 
3
  RUN useradd -m -u 1000 user
4
  USER user
 
1
+ # Use Python 3.12 slim image as base
2
+ FROM python:3.12-slim
3
 
4
  RUN useradd -m -u 1000 user
5
  USER user
main/logs/llm_api.log CHANGED
@@ -51,3 +51,24 @@
51
  2025-01-09 17:15:14,956 - llm_api - INFO - Loading model from local path: main/models/phi-4
52
  2025-01-09 17:15:14,965 - llm_api - ERROR - Failed to initialize generation model microsoft/phi-4: CUDA is required but not available for bitsandbytes. Please consider installing the multi-platform enabled version of bitsandbytes, which is currently a work in progress. Please check currently supported platforms and installation instructions at https://huggingface.co/docs/bitsandbytes/main/en/installation#multi-backend
53
  2025-01-09 17:15:14,965 - api_routes - ERROR - Error initializing model: CUDA is required but not available for bitsandbytes. Please consider installing the multi-platform enabled version of bitsandbytes, which is currently a work in progress. Please check currently supported platforms and installation instructions at https://huggingface.co/docs/bitsandbytes/main/en/installation#multi-backend
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
51
  2025-01-09 17:15:14,956 - llm_api - INFO - Loading model from local path: main/models/phi-4
52
  2025-01-09 17:15:14,965 - llm_api - ERROR - Failed to initialize generation model microsoft/phi-4: CUDA is required but not available for bitsandbytes. Please consider installing the multi-platform enabled version of bitsandbytes, which is currently a work in progress. Please check currently supported platforms and installation instructions at https://huggingface.co/docs/bitsandbytes/main/en/installation#multi-backend
53
  2025-01-09 17:15:14,965 - api_routes - ERROR - Error initializing model: CUDA is required but not available for bitsandbytes. Please consider installing the multi-platform enabled version of bitsandbytes, which is currently a work in progress. Please check currently supported platforms and installation instructions at https://huggingface.co/docs/bitsandbytes/main/en/installation#multi-backend
54
+ 2025-01-13 16:04:32,247 - hf_validation - WARNING - No .env file found. Fine if you're on Huggingface, but you need one to run locally on your PC.
55
+ 2025-01-13 16:04:32,247 - hf_validation - ERROR - No HF_TOKEN found in environment variables
56
+ 2025-01-13 16:04:32,247 - main - INFO - Starting LLM API server
57
+ 2025-01-13 16:04:32,248 - llm_api - INFO - Initializing LLM API
58
+ 2025-01-13 16:04:32,248 - llm_api - INFO - LLM API initialized successfully
59
+ 2025-01-13 16:04:32,248 - api_routes - INFO - Router initialized with LLM API instance
60
+ 2025-01-13 16:04:32,252 - main - INFO - FastAPI application created successfully
61
+ 2025-01-13 16:05:27,996 - api_routes - INFO - Received request to download model: microsoft/Phi-3.5-mini-instruct
62
+ 2025-01-13 16:05:27,996 - llm_api - INFO - Starting download of model: microsoft/Phi-3.5-mini-instruct
63
+ 2025-01-13 16:05:27,996 - llm_api - INFO - Enabling stdout logging for download
64
+ 2025-01-13 16:08:46,773 - llm_api - INFO - Disabling stdout logging
65
+ 2025-01-13 16:08:46,773 - llm_api - INFO - Saving model to main/models/Phi-3.5-mini-instruct
66
+ 2025-01-13 16:10:23,543 - llm_api - INFO - Successfully downloaded model: microsoft/Phi-3.5-mini-instruct
67
+ 2025-01-13 16:10:24,432 - api_routes - INFO - Successfully downloaded model: microsoft/Phi-3.5-mini-instruct
68
+ 2025-01-13 16:18:45,409 - api_routes - INFO - Received request to initialize model: microsoft/Phi-3.5-mini-instruct
69
+ 2025-01-13 16:18:45,409 - llm_api - INFO - Initializing generation model: microsoft/Phi-3.5-mini-instruct
70
+ 2025-01-13 16:18:45,412 - llm_api - INFO - Loading model from local path: main/models/Phi-3.5-mini-instruct
71
+ 2025-01-13 16:18:45,982 - llm_api - ERROR - Failed to initialize generation model microsoft/Phi-3.5-mini-instruct: Failed to import transformers.integrations.bitsandbytes because of the following error (look up to see its traceback):
72
+ Dynamo is not supported on Python 3.13+
73
+ 2025-01-13 16:18:45,982 - api_routes - ERROR - Error initializing model: Failed to import transformers.integrations.bitsandbytes because of the following error (look up to see its traceback):
74
+ Dynamo is not supported on Python 3.13+