Skip to main content

XGBoost runtime for MLServer

Project description

XGBoost runtime for MLServer

This package provides a MLServer runtime compatible with XGBoost.

Usage

You can install the runtime, alongside mlserver, as:

pip install mlserver mlserver-xgboost

For further information on how to use MLServer with XGBoost, you can check out this worked out example.

XGBoost Artifact Type

The XGBoost inference runtime will expect that your model is serialised via one of the following methods:

Extension Docs Example
*.json JSON Format booster.save_model("model.json")
*.ubj Binary JSON Format booster.save_model("model.ubj")
*.bst (Old) Binary Format booster.save_model("model.bst")
By default, the runtime will look for a file called `model.[json | ubj | bst]`.
However, this can be modified through the `parameters.uri` field of your
{class}`ModelSettings <mlserver.settings.ModelSettings>` config (see the
section on [Model Settings](../../docs/reference/model-settings.md) for more
details).

```{code-block} json
---
emphasize-lines: 3-5
---
{
  "name": "foo",
  "parameters": {
    "uri": "./my-own-model-filename.json"
  }
}
```

Content Types

If no content type is present on the request or metadata, the XGBoost runtime will try to decode the payload as a NumPy Array. To avoid this, either send a different content type explicitly, or define the correct one as part of your model's metadata.

Model Outputs

The XGBoost inference runtime exposes a number of outputs depending on the model type. These outputs match to the predict and predict_proba methods of the XGBoost model.

Output Returned By Default Availability
predict Available on all XGBoost models.
predict_proba Only available on non-regressor models (i.e. XGBClassifier models).

By default, the runtime will only return the output of predict. However, you are able to control which outputs you want back through the outputs field of your {class}InferenceRequest <mlserver.types.InferenceRequest> payload.

For example, to only return the model's predict_proba output, you could define a payload such as:

---
emphasize-lines: 10-12
---
{
  "inputs": [
    {
      "name": "my-input",
      "datatype": "INT32",
      "shape": [2, 2],
      "data": [1, 2, 3, 4]
    }
  ],
  "outputs": [
    { "name": "predict_proba" }
  ]
}

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

mlserver-xgboost-1.3.3rc1.tar.gz (7.9 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

mlserver_xgboost-1.3.3rc1-py3-none-any.whl (8.3 kB view details)

Uploaded Python 3

File details

Details for the file mlserver-xgboost-1.3.3rc1.tar.gz.

File metadata

  • Download URL: mlserver-xgboost-1.3.3rc1.tar.gz
  • Upload date:
  • Size: 7.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.8.16

File hashes

Hashes for mlserver-xgboost-1.3.3rc1.tar.gz
Algorithm Hash digest
SHA256 b0aedea70718f64f0c3034bcdbcc56d254286d5d3a970e996e116768c7db81cd
MD5 e5b1080e7cd76e4e9c17ab376f920b10
BLAKE2b-256 b10968f151c42880e62c80dbdb4046de33c2a245d7c132cb47d856c97b0f8c9c

See more details on using hashes here.

File details

Details for the file mlserver_xgboost-1.3.3rc1-py3-none-any.whl.

File metadata

File hashes

Hashes for mlserver_xgboost-1.3.3rc1-py3-none-any.whl
Algorithm Hash digest
SHA256 644535be1ef2e735c49dbb48705ab81fd469371b844fe33107a0074b4c24a4ac
MD5 df6bf357a0a8d6a0864a3e5b31d9fa30
BLAKE2b-256 1258052e739edf8d7aeffdb2f7d93c989733b1ad33c3886d145fc154554160e8

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page