Skip to content

Pull requests: NVIDIA/TensorRT-LLM

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Pull requests list

[None][fix] Clean up llmc licensing docs
#13700 opened May 2, 2026 by bmarimuthu-nv Collaborator Draft
1 task
[None][feat] Add bf16 trtllm moe through flashinfer.
#13689 opened May 1, 2026 by nv-guomingz Collaborator Draft
1 task done
[TRTLLM-12432][perf] ltx2: drop redundant pe all-gather in AV cross-attention
#13687 opened May 1, 2026 by luyiyun1021 Collaborator Loading…
1 task done
[None][chore] KV Cache Transceiver Profiling Configs
#13681 opened Apr 30, 2026 by ekou24 Collaborator Loading…
1 task
Eg/ad mla chunked prefill loop
#13677 opened Apr 30, 2026 by MrGeva Collaborator Draft
1 task
[None][fix] Fix GPT-OSS KV-aware router hashing
#13675 opened Apr 30, 2026 by SimengLiu-nv Collaborator Loading…
1 task done
[None][perf] Improve TRTLLM MoE autotune in DEP
#13667 opened Apr 30, 2026 by rosenrodt Collaborator Loading…
1 task done
[None][chore] Refactor attention forward context
#13662 opened Apr 30, 2026 by yuxianq Collaborator Draft
1 task done
ProTip! Type g i on any issue or pull request to go back to the issue listing page.