| library_name: transformers | |
| tags: | |
| - kv-cache | |
| - fp8 | |
| # Llama-3.1-8B-Instruct-QKV-Cache-FP8-Per-Tensor | |
| Empty model card placeholder. | |
| library_name: transformers | |
| tags: | |
| - kv-cache | |
| - fp8 | |
| # Llama-3.1-8B-Instruct-QKV-Cache-FP8-Per-Tensor | |
| Empty model card placeholder. | |