Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
Commit
Β·
b48ecb9
1
Parent(s):
e497915
fixing qwen model loading from HF dataset
Browse files- Dockerfile +11 -5
- ppo_logic.py +3 -2
Dockerfile
CHANGED
|
@@ -31,12 +31,18 @@ RUN mkdir -p /models/cbow && \
|
|
| 31 |
# RUN mkdir -p /models/qwen && \
|
| 32 |
# python3 -c "from transformers import AutoTokenizer; AutoTokenizer.from_pretrained('Qwen/Qwen3-0.6B-Base').save_pretrained('/models/qwen')"
|
| 33 |
# RUN python3 -c "from transformers import AutoModelForCausalLM; AutoModelForCausalLM.from_pretrained('Qwen/Qwen3-0.6B-Base').save_pretrained('/models/qwen')"
|
| 34 |
-
RUN mkdir -p /models/qwen && \
|
| 35 |
-
python3 -c "from transformers import AutoTokenizer; \
|
| 36 |
-
AutoTokenizer.from_pretrained('ClemSummer/qwen-model-cache', trust_remote_code=True).save_pretrained('/models/qwen')"
|
| 37 |
|
| 38 |
-
RUN
|
| 39 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 40 |
EXPOSE 7860
|
| 41 |
|
| 42 |
# Install curl if it's not already installed
|
|
|
|
| 31 |
# RUN mkdir -p /models/qwen && \
|
| 32 |
# python3 -c "from transformers import AutoTokenizer; AutoTokenizer.from_pretrained('Qwen/Qwen3-0.6B-Base').save_pretrained('/models/qwen')"
|
| 33 |
# RUN python3 -c "from transformers import AutoModelForCausalLM; AutoModelForCausalLM.from_pretrained('Qwen/Qwen3-0.6B-Base').save_pretrained('/models/qwen')"
|
|
|
|
|
|
|
|
|
|
| 34 |
|
| 35 |
+
# RUN mkdir -p /models/qwen && \
|
| 36 |
+
# python3 -c "from transformers import AutoTokenizer; \
|
| 37 |
+
# AutoTokenizer.from_pretrained('ClemSummer/qwen-model-cache', trust_remote_code=True).save_pretrained('/models/qwen')"
|
| 38 |
+
# RUN python3 -c "from transformers import AutoModelForCausalLM; \
|
| 39 |
+
# AutoModelForCausalLM.from_pretrained('ClemSummer/qwen-model-cache', trust_remote_code=True).save_pretrained('/models/qwen')"
|
| 40 |
+
|
| 41 |
+
# Install git & git-lfs, then clone your dataset repo into /models/qwen
|
| 42 |
+
RUN apt-get update && apt-get install -y git git-lfs && \
|
| 43 |
+
git lfs install && \
|
| 44 |
+
git clone https://huggingface.co/datasets/ClemSummer/qwen-model-cache /models/qwen
|
| 45 |
+
|
| 46 |
EXPOSE 7860
|
| 47 |
|
| 48 |
# Install curl if it's not already installed
|
ppo_logic.py
CHANGED
|
@@ -19,13 +19,14 @@ MAX_NEW_TOKENS = 256
|
|
| 19 |
# Load tokenizer and model
|
| 20 |
# -------------------------------
|
| 21 |
print("π Loading tokenizer and model...")
|
| 22 |
-
tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
|
| 23 |
tokenizer.pad_token = tokenizer.eos_token
|
| 24 |
|
| 25 |
base_model = AutoModelForCausalLM.from_pretrained(
|
| 26 |
MODEL_NAME,
|
| 27 |
torch_dtype=torch.float16,
|
| 28 |
-
device_map="auto"
|
|
|
|
| 29 |
)
|
| 30 |
|
| 31 |
model = PeftModel.from_pretrained(base_model, CHECKPOINT_DIR)
|
|
|
|
| 19 |
# Load tokenizer and model
|
| 20 |
# -------------------------------
|
| 21 |
print("π Loading tokenizer and model...")
|
| 22 |
+
tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME, trust_remote_code=True)
|
| 23 |
tokenizer.pad_token = tokenizer.eos_token
|
| 24 |
|
| 25 |
base_model = AutoModelForCausalLM.from_pretrained(
|
| 26 |
MODEL_NAME,
|
| 27 |
torch_dtype=torch.float16,
|
| 28 |
+
device_map="auto",
|
| 29 |
+
trust_remote_code=True
|
| 30 |
)
|
| 31 |
|
| 32 |
model = PeftModel.from_pretrained(base_model, CHECKPOINT_DIR)
|