inference-optimization/Qwen3-32B-FP8-dynamic-QKV-Cache-FP8-Per-Tensor 33B • Updated about 1 month ago • 2
inference-optimization/Qwen3-32B-FP8-dynamic-QKV-Cache-FP8-Per-Head 33B • Updated about 1 month ago • 1
inference-optimization/Llama-3.3-70B-Instruct-QKV-Cache-FP8-Per-Tensor 71B • Updated about 1 month ago • 2
inference-optimization/Llama-3.3-70B-Instruct-QKV-Cache-FP8-Per-Head 71B • Updated about 1 month ago • 2
inference-optimization/Llama-3.3-70B-Instruct-FP8-dynamic-QKV-Cache-FP8-Per-Tensor 71B • Updated about 1 month ago • 3
inference-optimization/Llama-3.3-70B-Instruct-FP8-dynamic-QKV-Cache-FP8-Per-Head 71B • Updated about 1 month ago • 1
inference-optimization/Llama-3.1-8B-Instruct-QKV-Cache-FP8-Per-Tensor 8B • Updated about 1 month ago • 19
inference-optimization/Llama-3.1-8B-Instruct-FP8-dynamic-QKV-Cache-FP8-Per-Tensor 8B • Updated about 1 month ago • 14