Update README.md
Browse files
README.md
CHANGED
|
@@ -94,7 +94,7 @@ Most quantizations use **256 calibration samples**. We use **2048+ diverse sampl
|
|
| 94 |
| Model | Params | Size | vs FP16 | vs FP8 | VRAM | Status |
|
| 95 |
|-------|--------|------|---------|--------|------|--------|
|
| 96 |
| [granite-34b-code-instruct-8k-Ultra-Hybrid](https://huggingface.co/TevunahAi/granite-34b-code-instruct-8k-Ultra-Hybrid) | 34B | 21.8GB | -68% | -37% | 20.4GB | ✅ Verified |
|
| 97 |
-
| [Llama-3.1-70B-Instruct-Ultra-Hybrid]
|
| 98 |
|
| 99 |
|
| 100 |
|
|
|
|
| 94 |
| Model | Params | Size | vs FP16 | vs FP8 | VRAM | Status |
|
| 95 |
|-------|--------|------|---------|--------|------|--------|
|
| 96 |
| [granite-34b-code-instruct-8k-Ultra-Hybrid](https://huggingface.co/TevunahAi/granite-34b-code-instruct-8k-Ultra-Hybrid) | 34B | 21.8GB | -68% | -37% | 20.4GB | ✅ Verified |
|
| 97 |
+
| [Llama-3.1-70B-Instruct-Ultra-Hybrid](https://huggingface.co/TevunahAi/Llama-3.1-70B-Instruct-Ultra-Hybrid) | 70B | 45.4GB | -68% | -37% | 48GB | ✅ Verified |
|
| 98 |
|
| 99 |
|
| 100 |
|