Skip to content

Conversation

@rupeng-liu
Copy link
Contributor

Thanks to @bythew3i (Jevin) for providing the insight for the optimization. This is to skip size computation during DMA wait.

I have ran both kernel's test and brought up e2e vllm server to test, will provide perf improvement numbers

Signed-off-by: Rupeng Liu rupengliu@meta.com

Description

Start with a short description of what the PR does and how this is a change from
the past.

The rest of the description includes relevant details and context, examples:

  • why is this change being made,
  • the problem being solved and any relevant context,
  • why this is a good solution,
  • some information about the specific implementation,
  • shortcomings of the solution and possible future improvements.

If the change fixes a bug or a Github issue, please include a link, e.g.,:
FIXES: b/123456
FIXES: #123456

Tests

Please describe how you tested this change, and include any instructions and/or
commands to reproduce.

Checklist

Before submitting this PR, please make sure:

  • I have performed a self-review of my code.
  • I have necessary comments in my code, particularly in hard-to-understand areas.
  • I have made or will make corresponding changes to any relevant documentation.

Signed-off-by: Rupeng Liu rupengliu@meta.com
@rupengliu-meta
Copy link
Contributor

around 5-10% throughput improvement

sem,
wait,
src=vmem_ref,
dst=vmem_ref.at[pl.ds(0, offset + bkv_sz_frm_new)],
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

wait only cares about size.

Please refer to https://github.com/vllm-project/tpu-inference/pull/718/files

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

thankjs Jevin, please check #1126

@rupengliu-meta
Copy link
Contributor

rupengliu-meta commented Nov 19, 2025

moved to a new github account, close this PR for now

@rupeng-liu rupeng-liu closed this Nov 19, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants