Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

append_kv_cache's documentation is out of date #616

Open
reyoung opened this issue Nov 19, 2024 · 1 comment
Open

append_kv_cache's documentation is out of date #616

reyoung opened this issue Nov 19, 2024 · 1 comment
Assignees

Comments

@reyoung
Copy link
Contributor

reyoung commented Nov 19, 2024

I noticed in the main branch, that the append_kv_cache does not take append_indptr any more. But the document reference append_indptr. It is a little bit confusing.

And I cannot find the batch indices documentation for a detailed situation.

  • Is this tensor [0, 0, 2, 3, 5, 5] a valid batch indices?
    • It happens when some prefilling sample's kv are fully cached in a prefilling batch.
    • I just gathered the input tensor before append_kv_cache when append_kv_cache takes append_indptr. Is this gathering unnecessary in the new API?
    • Can I specify batch_indices = -1 to skip some append_kv_cache ?
@zhyncs
Copy link
Member

zhyncs commented Nov 19, 2024

Hi @reyoung Recently, FlashInfer has been undergoing extensive refactoring, with tens of thousands of lines of code to be updated soon. The documentation hasn't been updated in time, and it will be once @yzh119 has the bandwidth.

@yzh119 yzh119 self-assigned this Nov 20, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants