Skip to content

chore: support getting the latest iteration status #3414

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
wants to merge 10 commits into
base: main
Choose a base branch
from

Conversation

pansicheng
Copy link
Contributor

This modification introduces iter_perf_stats_size to control the number of recent iter_perf_stats returned, improving metrics API response time.

@juney-nvidia juney-nvidia changed the title support getting the latest iteration status chore: support getting the latest iteration status Apr 9, 2025
@kaiyux kaiyux added Community want to contribute PRs initiated from Community Community Engagement help/insights needed from community labels Apr 10, 2025
@kaiyux
Copy link
Member

kaiyux commented Apr 15, 2025

/bot run

@tensorrt-cicd
Copy link
Collaborator

PR_Github #2343 [ run ] triggered by Bot

@tensorrt-cicd
Copy link
Collaborator

PR_Github #2343 [ run ] completed with state FAILURE
/LLM/main/L0_MergeRequest_PR pipeline #1685 completed with status: 'FAILURE'

@kaiyux
Copy link
Member

kaiyux commented Apr 15, 2025

Hi @pansicheng , the pipeline is failed on style check, can you help fix it following the guidance here? https://github.com/NVIDIA/TensorRT-LLM/blob/main/CONTRIBUTING.md#coding-style

@pansicheng pansicheng force-pushed the latest-metrics branch 2 times, most recently from 4944891 to d8bdf57 Compare April 17, 2025 01:53
@kaiyux
Copy link
Member

kaiyux commented Apr 17, 2025

/bot run

@tensorrt-cicd
Copy link
Collaborator

PR_Github #2571 [ run ] triggered by Bot

@tensorrt-cicd
Copy link
Collaborator

PR_Github #2571 [ run ] completed with state SUCCESS
/LLM/main/L0_MergeRequest_PR pipeline #1837 completed with status: 'FAILURE'

@qsang-nv
Copy link
Collaborator

/bot run

@kaiyux
Copy link
Member

kaiyux commented Apr 17, 2025

/bot run

@tensorrt-cicd
Copy link
Collaborator

PR_Github #2687 [ run ] triggered by Bot

@tensorrt-cicd
Copy link
Collaborator

PR_Github #2687 [ run ] completed with state SUCCESS
/LLM/main/L0_MergeRequest_PR pipeline #1917 completed with status: 'FAILURE'

Copy link
Collaborator

@Superjomn Superjomn left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM, thanks for the contribution!

@kaiyux
Copy link
Member

kaiyux commented Apr 22, 2025

/bot run

@tensorrt-cicd
Copy link
Collaborator

PR_Github #3046 [ run ] triggered by Bot

@tensorrt-cicd
Copy link
Collaborator

PR_Github #3046 [ run ] completed with state FAILURE
/LLM/main/L0_MergeRequest_PR pipeline #2129 completed with status: 'FAILURE'

@syuoni
Copy link
Collaborator

syuoni commented Apr 23, 2025

/bot run

@tensorrt-cicd
Copy link
Collaborator

PR_Github #3111 [ run ] triggered by Bot

@tensorrt-cicd
Copy link
Collaborator

PR_Github #3111 [ run ] completed with state SUCCESS
/LLM/main/L0_MergeRequest_PR pipeline #2168 completed with status: 'FAILURE'

@qsang-nv
Copy link
Collaborator

/bot run

@kaiyux
Copy link
Member

kaiyux commented May 1, 2025

/bot run

@tensorrt-cicd
Copy link
Collaborator

PR_Github #3930 [ run ] triggered by Bot

@tensorrt-cicd
Copy link
Collaborator

PR_Github #3930 [ run ] completed with state SUCCESS
/LLM/main/L0_MergeRequest_PR pipeline #2788 completed with status: 'FAILURE'

@poweiw
Copy link
Collaborator

poweiw commented Jun 5, 2025

@kaiyux what's the next step for this pr?

@poweiw poweiw added triaged Issue has been triaged by maintainers Performance TRTLLM model inference speed, throughput, efficiency. Latency, benchmarks, regressions, opts. labels Jun 5, 2025
@pansicheng pansicheng requested a review from a team as a code owner June 7, 2025 06:25
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Community Engagement help/insights needed from community Community want to contribute PRs initiated from Community Performance TRTLLM model inference speed, throughput, efficiency. Latency, benchmarks, regressions, opts. triaged Issue has been triaged by maintainers
Projects
None yet
Development

Successfully merging this pull request may close these issues.

7 participants