Skip to content

Latest commit

 

History

History
63 lines (43 loc) · 1.1 KB

README.md

File metadata and controls

63 lines (43 loc) · 1.1 KB

Sentence Embedding Serving Server

Getting Started

Run Docker

docker run -p 8080:8080 -it ghcr.io/codelibs/embedding-api:1.0.1

Run Docker with Model Name

docker run -p 8080:8080 -e MODEL_NAME=intfloat/multilingual-e5-large -it ghcr.io/codelibs/embedding-api:1.0.1

Use Model Cache

docker run -v ./model:/code/model -p 8080:8080 -it ghcr.io/codelibs/embedding-api:1.0.1

Run Docker with GPU

docker run --gpus all -p 8080:8080 -it ghcr.io/codelibs/embedding-api:1.0.1.cuda11

Request

Embedding API

curl -s -H "Content-Type:application/json" -XPOST localhost:8080/encode -d '
{
  "sentences": [
    "This framework generates embeddings for each input sentence",
    "Sentences are passed as a list of string.",
    "The quick brown fox jumps over the lazy dog."
  ]
}'

Information API

curl -s -H "Content-Type:application/json" -XGET localhost:8080/

Ping API

curl -s -H "Content-Type:application/json" -XGET localhost:8080/ping

Build

Build Docker

docker build --rm -t ghcr.io/codelibs/embedding-api:1.0.1 .