-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Not Sure How to Apply This to Docker ollama/ollama:rocm #2
Comments
|
I am trying to get Perplexica, a local open-source version of Perplexity.ai, running and using an ollama that accesses rocm. I have gotten perplexica (https://github.com/ItzCrazyKns/Perplexica) working using a regular ollama in a docker but it's painfully slow. My system has rocm and supports Vulkan and I have Ryzen 780M. I am trying to get ollama set up in a docker to use rocm. I think when I wrote this issue before, I didn't understand that you were validating programs as working correctly with rocm. Perhaps I should have titled the issue "Please help me." https://github.com/avnigashi/ollama-rocm-docker is one of the repos that isn't seeming to work for me. https://github.com/hqnicolas/OllamaDockerCasaOs is another repo I am not entirely understanding. I don't understand why casaos is part of this equation. This version runs, but generates errors in the docker terminal "Error: unknown flag: --device" and I'm not sure if it's running correctly. my experience level is very low, so it's harder for me to understand things than a more experienced user |
Hi @amd1890 , This repo is my practice to using ROCm to power other AI projects and want to back to OSS comunity to sharing my experience. For your question, I never use Perplexica with ROCm before. Base on your situation, you many try set up Ollama with AMD 780M first refer to https://github.com/alexhegit/Playing-with-ROCm/blob/main/inference/LLM/Run_Ollama_with_AMD_iGPU780M-QuickStart.md and then to access Preplexica with Ollama+ROCm. I also add it as my wishlist and will find time to try it later. GOOD LUCK! |
I followed those instructions and could get ollama to run with an extremely high token/s count, indicating using the GPU. I am still not sure how to integrate that into perplexica. https://github.com/ItzCrazyKns/Perplexica Step 4 for the docker installation is: Rename the sample.config.toml file to config.toml. For Docker setups, you need only fill in the following fields:
I do not understand if this has to be a docker ollama. I think it does have to be? This is so hard I may end up giving up. I may not be smart enough to figure this out. |
I believe that you have set up the ollama over AMD GPU by following my instructions and you can run it like Then you can follow the instructions of https://github.com/ItzCrazyKns/Perplexica to run Perplexica. Here is my trying run Pefplexica with docker at the same machine which ollama serving.
Then modify the config.toml for setting of OLLAMA as bellow,
Start the Perplexica docker (pull, build, start...)
Wait a few minutes for the setup to complete. You can access Perplexica at http://localhost:3000/ in your web browser. I could open it by the web browser but it seems that it not running successfully. You can get my issue report from |
I've been trying to get this to work in docker which is using ollama and is painfully slow
It doesn't recognize the GPU and I don't know how to implement the commands inside the docker image or get the image to recognize the GPU.
Anyone have suggestions?
The text was updated successfully, but these errors were encountered: