Replies: 1 comment 2 replies
-
sorry I'm not the expert on severing stuff. Could you check the problem on llama.cpp, where it has already integrated embedding models in recent PR based on this one ? |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
@xyzhang626 is there any possibility of running the embedding model behind a rest API server instead of socket-based connection?
Beta Was this translation helpful? Give feedback.
All reactions