Skip to content

ggml-org/hf-inference-endpoints

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

13 Commits
 
 
 
 

Repository files navigation

Inference endpoints for Hugging Face

This repo contains Docker containers that can be used to deploy ggml-based inference endpoints at:

https://ui.endpoints.huggingface.co

Instructions

During dedicated endpoint creation, select custom container type like this:

image

note: the LLAMACPP_ARGS environment variable is a temporary mechanism to pass custom arguments to llama-server

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Dockerfile 100.0%