Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

RKNN-LLM Context length expansion #93

Open
vincenzodentamaro opened this issue Sep 9, 2024 · 0 comments
Open

RKNN-LLM Context length expansion #93

vincenzodentamaro opened this issue Sep 9, 2024 · 0 comments

Comments

@vincenzodentamaro
Copy link

vincenzodentamaro commented Sep 9, 2024

Because of the lack of embeddings extraction to create a vector space database, I need to increase the context length up to 32k or more if possible.
I see that some models such as Qwen2 support up to 128k context. How can I increase it.

Should I have to update the rknn_api.h:

#define RKNN_MAX_DIMS 32 /* increased maximum dimension of tensor. /
#define RKNN_MAX_NUM_CHANNEL 20 /
increased maximum channel number of input tensor. /
#define RKNN_MAX_NAME_LEN 512 /
increased maximum name length of tensor. /
#define RKNN_MAX_DYNAMIC_SHAPE_NUM 32768 /
increased maximum number of dynamic shape for each input. */

or is it enough to set param.max_context_len = 32768; ?

@airockchip @waydong

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant