Update app.py
Browse files
app.py
CHANGED
|
@@ -30,7 +30,7 @@ from transformers import AutoModelForCausalLM, AutoTokenizer
|
|
| 30 |
model_id = "CohereForAI/c4ai-command-r-plus-4bit"
|
| 31 |
|
| 32 |
|
| 33 |
-
tokenizer = AutoTokenizer.from_pretrained(model_id, token= token,
|
| 34 |
|
| 35 |
model = AutoModelForCausalLM.from_pretrained(model_id, token= token,
|
| 36 |
# torch_dtype=torch.bfloat16,
|
|
@@ -38,7 +38,7 @@ model = AutoModelForCausalLM.from_pretrained(model_id, token= token,
|
|
| 38 |
# low_cpu_mem_usage=True,
|
| 39 |
# llm_int8_enable_fp32_cpu_offload=True,
|
| 40 |
device_map="cuda",
|
| 41 |
-
|
| 42 |
)
|
| 43 |
|
| 44 |
|
|
|
|
| 30 |
model_id = "CohereForAI/c4ai-command-r-plus-4bit"
|
| 31 |
|
| 32 |
|
| 33 |
+
tokenizer = AutoTokenizer.from_pretrained(model_id, token= token,)
|
| 34 |
|
| 35 |
model = AutoModelForCausalLM.from_pretrained(model_id, token= token,
|
| 36 |
# torch_dtype=torch.bfloat16,
|
|
|
|
| 38 |
# low_cpu_mem_usage=True,
|
| 39 |
# llm_int8_enable_fp32_cpu_offload=True,
|
| 40 |
device_map="cuda",
|
| 41 |
+
|
| 42 |
)
|
| 43 |
|
| 44 |
|