Skip to content

Adding cuda kernel (optimized for sm80) for block-wise 4b quantized float 16 GEMM. #21834

Adding cuda kernel (optimized for sm80) for block-wise 4b quantized float 16 GEMM.

Adding cuda kernel (optimized for sm80) for block-wise 4b quantized float 16 GEMM. #21834

Triggered via pull request February 28, 2024 23:40
Status Success
Total duration 5m 19s
Artifacts

codeql.yml

on: pull_request
Matrix: Analyze
Fit to window
Zoom out
Zoom in