Update README.md
Browse files
README.md
CHANGED
|
@@ -23,7 +23,10 @@ Evaluated against the benchmark test: [RAG-Instruct-Benchmark-Tester](https://
|
|
| 23 |
--Summarization Quality (1-5): 4 (Above Average)
|
| 24 |
--Hallucinations: No hallucinations observed in test runs.
|
| 25 |
|
| 26 |
-
For test run results (and good indicator of target use cases), please see the files ("core_rag_test" and "answer_sheet" in this repo).
|
|
|
|
|
|
|
|
|
|
| 27 |
|
| 28 |
### Model Description
|
| 29 |
|
|
@@ -79,8 +82,7 @@ Load in your favorite GGUF inference engine, or try with llmware as follows:
|
|
| 79 |
|
| 80 |
# to load the model and make a basic inference
|
| 81 |
model = ModelCatalog().load_model("llmware/bling-phi-3-gguf", temperature=0.0, sample=False)
|
| 82 |
-
response = model.
|
| 83 |
-
|
| 84 |
|
| 85 |
Details on the prompt wrapper and other configurations are on the config.json file in the files repository.
|
| 86 |
|
|
|
|
| 23 |
--Summarization Quality (1-5): 4 (Above Average)
|
| 24 |
--Hallucinations: No hallucinations observed in test runs.
|
| 25 |
|
| 26 |
+
For test run results (and good indicator of target use cases), please see the files ("core_rag_test" and "answer_sheet" in this repo).
|
| 27 |
+
|
| 28 |
+
Note: compare results with [bling-phi-2](https://www.huggingface.co/llmware/bling-phi-2-v0), and [dragon-mistral-7b](https://www.huggingface.com/llmware/dragon-mistral-7b-v0).
|
| 29 |
+
|
| 30 |
|
| 31 |
### Model Description
|
| 32 |
|
|
|
|
| 82 |
|
| 83 |
# to load the model and make a basic inference
|
| 84 |
model = ModelCatalog().load_model("llmware/bling-phi-3-gguf", temperature=0.0, sample=False)
|
| 85 |
+
response = model.inference(query, add_context=text_sample)
|
|
|
|
| 86 |
|
| 87 |
Details on the prompt wrapper and other configurations are on the config.json file in the files repository.
|
| 88 |
|