Update app.py
Browse files
app.py
CHANGED
|
@@ -120,8 +120,8 @@ import transformers
|
|
| 120 |
# model = load_checkpoint_and_dispatch(model, model_id, device_map=device_map, no_split_module_classes=["GPTJBlock"])
|
| 121 |
# model.half()
|
| 122 |
|
| 123 |
-
MODEL_ID = "deepseek-ai/DeepSeek-R1-Distill-Qwen-14B"
|
| 124 |
-
|
| 125 |
|
| 126 |
CHAT_TEMPLATE = "َAuto"
|
| 127 |
MODEL_NAME = MODEL_ID.split("/")[-1]
|
|
@@ -134,7 +134,7 @@ quantization_config = BitsAndBytesConfig(
|
|
| 134 |
bnb_4bit_compute_dtype=torch.bfloat16
|
| 135 |
)
|
| 136 |
tokenizer = AutoTokenizer.from_pretrained(MODEL_ID)
|
| 137 |
-
model =
|
| 138 |
MODEL_ID,
|
| 139 |
device_map="auto",
|
| 140 |
low_cpu_mem_usage=True,
|
|
|
|
| 120 |
# model = load_checkpoint_and_dispatch(model, model_id, device_map=device_map, no_split_module_classes=["GPTJBlock"])
|
| 121 |
# model.half()
|
| 122 |
|
| 123 |
+
# MODEL_ID = "deepseek-ai/DeepSeek-R1-Distill-Qwen-14B"
|
| 124 |
+
MODEL_ID = "microsoft/phi-4"
|
| 125 |
|
| 126 |
CHAT_TEMPLATE = "َAuto"
|
| 127 |
MODEL_NAME = MODEL_ID.split("/")[-1]
|
|
|
|
| 134 |
bnb_4bit_compute_dtype=torch.bfloat16
|
| 135 |
)
|
| 136 |
tokenizer = AutoTokenizer.from_pretrained(MODEL_ID)
|
| 137 |
+
model = AutoModelForCausalLM.from_pretrained(
|
| 138 |
MODEL_ID,
|
| 139 |
device_map="auto",
|
| 140 |
low_cpu_mem_usage=True,
|