Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Predicting memory usage - Memory Access Violation #952

Open
AgentSmithers opened this issue Oct 16, 2024 · 2 comments
Open

Predicting memory usage - Memory Access Violation #952

AgentSmithers opened this issue Oct 16, 2024 · 2 comments

Comments

@AgentSmithers
Copy link

Description

When running InferAsync with 16Gb of ram I hit a peek of 100% memory usage. I added memory to resolve the issue but was wondering if there is a way to predict the memory usage by calculating the size of the token input to cancel and give feedback that additional ram is required instead of triggering a memory corruption/violation error. Anyone know if this calculation is possible?

@martindevans
Copy link
Member

I'd recommend asking this upstream in llama.cpp, whatever they say will apply to LLamaSharp too.

@AgentSmithers
Copy link
Author

Thanks for the feedback. I'll give it a go with them.

@AgentSmithers AgentSmithers changed the title Predicting memory useage - Memory Access Violtion Predicting memory usage - Memory Access Violation Oct 16, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants