#nlp #serving #model #api-server #grpc #http #grpc-api

app fasttext-serving

fastText model serving API server

34 releases

0.7.0 Jan 8, 2023
0.6.2 Mar 21, 2021
0.6.1 Dec 22, 2020
0.6.0-alpha.7 Sep 14, 2020
0.2.1 Feb 28, 2018

#469 in HTTP server

Download history 21/week @ 2024-02-16 55/week @ 2024-02-23 3/week @ 2024-03-01 256/week @ 2024-03-29

256 downloads per month

MIT license

30KB
641 lines

fasttext-serving

GitHub Actions Crates.io Docker Pulls

fastText model serving service

Installation

You can download prebuilt binary from GitHub releases, or install it using Cargo:

cargo install fasttext-serving

Using Docker:

docker pull messense/fasttext-serving

Usage

$ fasttext-serving --help

USAGE:
    fasttext-serving [OPTIONS] --model <model>

FLAGS:
        --grpc       Serving gRPC API instead of HTTP API
    -h, --help       Prints help information
    -V, --version    Prints version information

OPTIONS:
    -a, --address <address>    Listen address [default: 127.0.0.1]
    -m, --model <model>        Model path
    -p, --port <port>          Listen port [default: 8000]
    -w, --workers <workers>    Worker thread count, defaults to CPU count

Serve HTTP REST API

HTTP API endpoint:

POST /predict

Post data should be JSON array of string, for example ["abc", "def"]

CURL example:

$ curl -X POST -H 'Content-Type: application/json' \
     --data "[\"Which baking dish is best to bake a banana bread?\", \"Why not put knives in the dishwasher?\"]" \
     'http://localhost:8000/predict'
[[["baking"],[0.7152988]],[["equipment"],[0.73479545]]]

Serve gRPC API

Run the command with --grpc to serve gRPC API instead of HTTP REST API.

Please refer to gRPC Python client documentation here.

License

This work is released under the MIT license. A copy of the license is provided in the LICENSE file.

Dependencies

~10–25MB
~254K SLoC