Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How do I use it in vllm deployment #3

Open
jchang98 opened this issue Mar 5, 2024 · 6 comments
Open

How do I use it in vllm deployment #3

jchang98 opened this issue Mar 5, 2024 · 6 comments

Comments

@jchang98
Copy link

jchang98 commented Mar 5, 2024

How can I use this approach in vllm deployment without training,can you give me a specific example. thx

@ChenxinAn-fdu
Copy link
Contributor

Thank you for bringing this to our attention. Unfortunately, the current version of vLLM does not support the return of attention scores. However, we are pleased to inform you that this functionality is planned in the next release of the software.

In the meantime, we are working diligently to implement paged attention—the key feature of vLLM—as well as Flash decoding. These enhancements aim to accelerate the generation process and decrease the GPU memory of the KV cache.

we appreciate your patience while we work on these developments.
Stay tuned for updates.

@jchang98
Copy link
Author

jchang98 commented Mar 6, 2024

@ChenxinAn-fdu OK, thanks for your response

@ChenxinAn-fdu
Copy link
Contributor

I have pushed the code for flash decoding and it significantly decreases the memory consumption for decoding with KV-cache. It may be helpful for you.

@skyshine102
Copy link

looking forward to the support in vllm!

@Shuai-Xie
Copy link

@ChenxinAn-fdu Dose vllm support DCA now? We'd like to use this feature in the deployment.

@ChenxinAn-fdu
Copy link
Contributor

@Shuai-Xie Hi, I left an issue in their official repo, but it seems that the current version of vllm only supports returning the output tensor without softmax_lse. We plan to implement it ourselves.

If you do not need continual batching, the current repo has implemented flash_decoding. You can use it for some preliminary experiments.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants