cecilia-uu
cecilia-uu
commited on
Commit
·
d862dfe
1
Parent(s):
cd3f5c7
Adding the Minimax model (#1009)
Browse files### What problem does this PR solve?
Added support for MiniMax LLM
### Type of change
- [x] New Feature (non-breaking change which adds functionality)
---------
Co-authored-by: cecilia-uu <[email protected]>
- api/db/init_data.py +48 -0
- rag/llm/__init__.py +2 -1
- rag/llm/chat_model.py +8 -0
api/db/init_data.py
CHANGED
|
@@ -152,6 +152,11 @@ factory_infos = [{
|
|
| 152 |
"logo": "",
|
| 153 |
"tags": "TEXT EMBEDDING, TEXT RE-RANK",
|
| 154 |
"status": "1",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 155 |
}
|
| 156 |
# {
|
| 157 |
# "name": "文心一言",
|
|
@@ -536,6 +541,49 @@ def init_llm_factory():
|
|
| 536 |
"max_tokens": 2048,
|
| 537 |
"model_type": LLMType.RERANK.value
|
| 538 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 539 |
]
|
| 540 |
for info in factory_infos:
|
| 541 |
try:
|
|
|
|
| 152 |
"logo": "",
|
| 153 |
"tags": "TEXT EMBEDDING, TEXT RE-RANK",
|
| 154 |
"status": "1",
|
| 155 |
+
},{
|
| 156 |
+
"name": "Minimax",
|
| 157 |
+
"logo": "",
|
| 158 |
+
"tags": "LLM,TEXT EMBEDDING",
|
| 159 |
+
"status": "1",
|
| 160 |
}
|
| 161 |
# {
|
| 162 |
# "name": "文心一言",
|
|
|
|
| 541 |
"max_tokens": 2048,
|
| 542 |
"model_type": LLMType.RERANK.value
|
| 543 |
},
|
| 544 |
+
# ------------------------ Minimax -----------------------
|
| 545 |
+
{
|
| 546 |
+
"fid": factory_infos[13]["name"],
|
| 547 |
+
"llm_name": "abab6.5-chat",
|
| 548 |
+
"tags": "LLM,CHAT,8k",
|
| 549 |
+
"max_tokens": 8192,
|
| 550 |
+
"model_type": LLMType.CHAT.value
|
| 551 |
+
},
|
| 552 |
+
{
|
| 553 |
+
"fid": factory_infos[13]["name"],
|
| 554 |
+
"llm_name": "abab6.5s-chat",
|
| 555 |
+
"tags": "LLM,CHAT,245k",
|
| 556 |
+
"max_tokens": 245760,
|
| 557 |
+
"model_type": LLMType.CHAT.value
|
| 558 |
+
},
|
| 559 |
+
{
|
| 560 |
+
"fid": factory_infos[13]["name"],
|
| 561 |
+
"llm_name": "abab6.5t-chat",
|
| 562 |
+
"tags": "LLM,CHAT,8k",
|
| 563 |
+
"max_tokens": 8192,
|
| 564 |
+
"model_type": LLMType.CHAT.value
|
| 565 |
+
},
|
| 566 |
+
{
|
| 567 |
+
"fid": factory_infos[13]["name"],
|
| 568 |
+
"llm_name": "abab6.5g-chat",
|
| 569 |
+
"tags": "LLM,CHAT,8k",
|
| 570 |
+
"max_tokens": 8192,
|
| 571 |
+
"model_type": LLMType.CHAT.value
|
| 572 |
+
},
|
| 573 |
+
{
|
| 574 |
+
"fid": factory_infos[13]["name"],
|
| 575 |
+
"llm_name": "abab5.5-chat",
|
| 576 |
+
"tags": "LLM,CHAT,16k",
|
| 577 |
+
"max_tokens": 16384,
|
| 578 |
+
"model_type": LLMType.CHAT.value
|
| 579 |
+
},
|
| 580 |
+
{
|
| 581 |
+
"fid": factory_infos[13]["name"],
|
| 582 |
+
"llm_name": "abab5.5s-chat",
|
| 583 |
+
"tags": "LLM,CHAT,8k",
|
| 584 |
+
"max_tokens": 8192,
|
| 585 |
+
"model_type": LLMType.CHAT.value
|
| 586 |
+
},
|
| 587 |
]
|
| 588 |
for info in factory_infos:
|
| 589 |
try:
|
rag/llm/__init__.py
CHANGED
|
@@ -51,7 +51,8 @@ ChatModel = {
|
|
| 51 |
"Xinference": XinferenceChat,
|
| 52 |
"Moonshot": MoonshotChat,
|
| 53 |
"DeepSeek": DeepSeekChat,
|
| 54 |
-
"BaiChuan": BaiChuanChat
|
|
|
|
| 55 |
}
|
| 56 |
|
| 57 |
|
|
|
|
| 51 |
"Xinference": XinferenceChat,
|
| 52 |
"Moonshot": MoonshotChat,
|
| 53 |
"DeepSeek": DeepSeekChat,
|
| 54 |
+
"BaiChuan": BaiChuanChat,
|
| 55 |
+
"MiniMax": MiniMaxChat
|
| 56 |
}
|
| 57 |
|
| 58 |
|
rag/llm/chat_model.py
CHANGED
|
@@ -464,3 +464,11 @@ class VolcEngineChat(Base):
|
|
| 464 |
except Exception as e:
|
| 465 |
yield ans + "\n**ERROR**: " + str(e)
|
| 466 |
yield tk_count
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 464 |
except Exception as e:
|
| 465 |
yield ans + "\n**ERROR**: " + str(e)
|
| 466 |
yield tk_count
|
| 467 |
+
|
| 468 |
+
|
| 469 |
+
class MiniMaxChat(Base):
|
| 470 |
+
def __init__(self, key, model_name="abab6.5s-chat",
|
| 471 |
+
base_url="https://api.minimax.chat/v1/text/chatcompletion_v2"):
|
| 472 |
+
if not base_url:
|
| 473 |
+
base_url="https://api.minimax.chat/v1/text/chatcompletion_v2"
|
| 474 |
+
super().__init__(key, model_name, base_url)
|