Update app.py
Browse files
app.py
CHANGED
|
@@ -31,7 +31,7 @@ from transformers import AutoModelForCausalLM, AutoTokenizer
|
|
| 31 |
|
| 32 |
# model_id = "mistralai/Mistral-7B-v0.3"
|
| 33 |
|
| 34 |
-
model_id = "CohereForAI/aya-23-
|
| 35 |
|
| 36 |
|
| 37 |
tokenizer = AutoTokenizer.from_pretrained(
|
|
@@ -46,9 +46,9 @@ with init_empty_weights():
|
|
| 46 |
torch_dtype=torch.float16,
|
| 47 |
# torch_dtype=torch.fl,
|
| 48 |
# attn_implementation="flash_attention_2",
|
| 49 |
-
|
| 50 |
# llm_int8_enable_fp32_cpu_offload=True,
|
| 51 |
-
|
| 52 |
|
| 53 |
)
|
| 54 |
|
|
|
|
| 31 |
|
| 32 |
# model_id = "mistralai/Mistral-7B-v0.3"
|
| 33 |
|
| 34 |
+
model_id = "CohereForAI/aya-23-8B"
|
| 35 |
|
| 36 |
|
| 37 |
tokenizer = AutoTokenizer.from_pretrained(
|
|
|
|
| 46 |
torch_dtype=torch.float16,
|
| 47 |
# torch_dtype=torch.fl,
|
| 48 |
# attn_implementation="flash_attention_2",
|
| 49 |
+
low_cpu_mem_usage=True,
|
| 50 |
# llm_int8_enable_fp32_cpu_offload=True,
|
| 51 |
+
device_map='cuda',
|
| 52 |
|
| 53 |
)
|
| 54 |
|