Skip to content

Conversation

@cjluo-nv
Copy link
Collaborator

@cjluo-nv cjluo-nv commented Nov 4, 2025

What does this PR do?

Type of change: ? Bug fix

Overview: ?

  1. Remove kv_cache_config in the generate API. It's no longer used in the code as well. We just estimate KV cache usage from other parameters
  2. Add max_seq_len in the generate API to better estimate the real KV cache usage.
  3. Assume default lm_eval max input sequence length to be 4096

@cjluo-nv cjluo-nv requested review from a team as code owners November 4, 2025 07:01
@cjluo-nv cjluo-nv changed the title Update LLM generate API for modelopt LLM eval [NVBUG: 5617733] Update LLM generate API for modelopt LLM eval Nov 4, 2025
@codecov
Copy link

codecov bot commented Nov 4, 2025

Codecov Report

✅ All modified and coverable lines are covered by tests.
✅ Project coverage is 73.43%. Comparing base (009bd1a) to head (997a111).
⚠️ Report is 3 commits behind head on main.

Additional details and impacted files
@@           Coverage Diff           @@
##             main     #498   +/-   ##
=======================================
  Coverage   73.43%   73.43%           
=======================================
  Files         180      180           
  Lines       18149    18149           
=======================================
  Hits        13328    13328           
  Misses       4821     4821           

☔ View full report in Codecov by Sentry.
📢 Have feedback on the report? Share it here.

🚀 New features to boost your workflow:
  • ❄️ Test Analytics: Detect flaky tests, report on failures, and find test suite problems.

@cjluo-nv cjluo-nv merged commit ce8ce22 into main Nov 4, 2025
26 checks passed
@cjluo-nv cjluo-nv deleted the chenjiel/update_generate branch November 4, 2025 19:08
kevalmorabia97 pushed a commit that referenced this pull request Nov 4, 2025
## What does this PR do?

**Type of change:** ? Bug fix

**Overview:** ?

1) Remove kv_cache_config in the generate API. It's no longer used in
the code as well. We just estimate KV cache usage from other parameters
2) Add max_seq_len in the generate API to better estimate the real KV
cache usage.
3) Assume default lm_eval max input sequence length to be 4096

Signed-off-by: Chenjie Luo <[email protected]>
mxinO pushed a commit that referenced this pull request Nov 11, 2025
## What does this PR do?

**Type of change:** ? Bug fix

**Overview:** ?

1) Remove kv_cache_config in the generate API. It's no longer used in
the code as well. We just estimate KV cache usage from other parameters
2) Add max_seq_len in the generate API to better estimate the real KV
cache usage.
3) Assume default lm_eval max input sequence length to be 4096

Signed-off-by: Chenjie Luo <[email protected]>
Signed-off-by: mxin <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants