docker build . --platform linux/amd64 -t huggingface/tei
cargo install --path router -F candle -F metal
model=intfloat/multilingual-e5-large
revision=main
volume=$PWD/data
text-embeddings-router --model-id $model --revision $revision --port 8081
curl 127.0.0.1:8081/embed \
-X POST \
-d '{"inputs":"What is Deep Learning?"}' \
-H 'Content-Type: application/json'
model=BAAI/bge-large-en-v1.5
revision=refs/pr/5
curl 127.0.0.1:8081/rerank \
-X POST \
-d '{"query":"What is Deep Learning?", "texts": ["Deep Learning is not...", "Deep learning is..."]}' \
-H 'Content-Type: application/json'
curl 127.0.0.1:8081/predict \
-X POST \
-d '{"inputs":"I like you."}' \
-H 'Content-Type: application/json'