Skip to content

Actions: NVIDIA/TensorRT-LLM

All workflows

Actions

Loading...
Loading

Showing runs from all workflows
662 workflow runs
662 workflow runs

Filter by Event

Filter by Status

Filter by Branch

Filter by Actor

qserve is slower then awq int4 for llama2-7b on H100
auto-assign #20: Issue #2509 labeled by hello-11
December 10, 2024 06:48 2s
December 10, 2024 06:48 2s
Failed to install by poetry add
Blossom-CI #98: Issue comment #2515 (comment) created by hello-11
December 10, 2024 06:41 4s
December 10, 2024 06:41 4s
int8 slower than bf16 on A100
auto-assign #19: Issue #2553 labeled by nv-guomingz
December 10, 2024 06:32 45s
December 10, 2024 06:32 45s
int8 slower than bf16 on A100
auto-assign #18: Issue #2553 labeled by nv-guomingz
December 10, 2024 06:21 55s
December 10, 2024 06:21 55s
How to extract the hidden_states before output_ids during the inference process.
Blossom-CI #97: Issue comment #2499 (comment) created by hello-11
December 10, 2024 06:12 4s
December 10, 2024 06:12 4s
Close inactive issues
Close inactive issues #353: Scheduled
December 10, 2024 06:08 16s main
December 10, 2024 06:08 16s
lora doesn't work when kv_cache is disabled
auto-assign #16: Issue #2543 labeled by nv-guomingz
December 10, 2024 06:01 41s
December 10, 2024 06:01 41s
[bug] Medusa example fails with vicuna 33B
auto-assign #15: Issue #2478 labeled by nv-guomingz
December 10, 2024 05:52 42s
December 10, 2024 05:52 42s
[QST] How to get the prefill latency and TPOT resepectly when using C++ runtime
auto-assign #14: Issue #2500 labeled by hello-11
December 10, 2024 05:51 1s
December 10, 2024 05:51 1s
[QST] How to get the prefill latency and TPOT resepectly when using C++ runtime
Blossom-CI #96: Issue comment #2500 (comment) created by hello-11
December 10, 2024 05:51 4s
December 10, 2024 05:51 4s
Close inactive issues
Close inactive issues #352: Scheduled
December 10, 2024 05:06 20s main
December 10, 2024 05:06 20s
[Question] Running custom Encoder Decoder model
Blossom-CI #95: Issue comment #2491 (comment) created by hello-11
December 10, 2024 04:56 4s
December 10, 2024 04:56 4s
Close inactive issues
Close inactive issues #351: Scheduled
December 10, 2024 04:07 15s main
December 10, 2024 04:07 15s
int4 not faster than fp16 and fp8
Blossom-CI #94: Issue comment #2487 (comment) created by Tracin
December 10, 2024 03:25 5s
December 10, 2024 03:25 5s
Issues with installing on Windows
Blossom-CI #93: Issue comment #2489 (comment) created by hello-11
December 10, 2024 03:25 6s
December 10, 2024 03:25 6s
December 10, 2024 03:21 3s
Close inactive issues
Close inactive issues #350: Scheduled
December 10, 2024 03:21 15s main
December 10, 2024 03:21 15s
int4 not faster than fp16 and fp8
auto-assign #12: Issue #2487 labeled by hello-11
December 10, 2024 03:20 3s
December 10, 2024 03:20 3s
How to install tensorrt-llm in python3.11?
Blossom-CI #92: Issue comment #2481 (comment) created by hello-11
December 10, 2024 02:57 4s
December 10, 2024 02:57 4s
Close inactive issues
Close inactive issues #349: Scheduled
December 10, 2024 02:33 14s main
December 10, 2024 02:33 14s
Close inactive issues
Close inactive issues #348: Scheduled
December 10, 2024 01:34 16s main
December 10, 2024 01:34 16s
Close inactive issues
Close inactive issues #347: Scheduled
December 10, 2024 00:27 19s main
December 10, 2024 00:27 19s
Close inactive issues
Close inactive issues #346: Scheduled
December 9, 2024 23:06 15s main
December 9, 2024 23:06 15s
int8 slower than bf16 on A100
auto-assign #11: Issue #2553 labeled by ShuaiShao93
December 9, 2024 23:01 3s
December 9, 2024 23:01 3s