Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

fix(ci): Changes for master for 1.6.2.rc1 #1972

Merged
merged 35 commits into from
Nov 26, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
35 commits
Select commit Hold shift + click to select a range
c2dbbb9
Update Changelog (#1905)
github-actions[bot] Sep 10, 2024
14a37d8
build(deps): bump mlflow from 2.16.0 to 2.16.2 in /runtimes/mlflow (#…
dependabot[bot] Sep 23, 2024
0bea0ee
build(deps): bump optimum from 1.21.4 to 1.22.0 in /runtimes/huggingf…
dependabot[bot] Sep 23, 2024
8a1477c
build(deps): bump scikit-learn from 1.5.1 to 1.5.2 in /runtimes/sklea…
dependabot[bot] Sep 23, 2024
70f17ef
build(deps): bump pandas from 2.2.2 to 2.2.3 in /runtimes/lightgbm (#…
dependabot[bot] Sep 23, 2024
4a595c2
build(deps-dev): bump pytorch-lightning from 2.3.1 to 2.4.0 (#1900)
dependabot[bot] Sep 23, 2024
42d2ca1
build(deps-dev): bump torchmetrics in /runtimes/mlflow (#1909)
dependabot[bot] Sep 24, 2024
1627842
build(deps-dev): bump datamodel-code-generator from 0.25.5 to 0.26.0 …
dependabot[bot] Sep 24, 2024
35f975b
add docs for gitbook (#1919)
sakoush Oct 7, 2024
83a9b48
build(deps): bump optimum from 1.22.0 to 1.23.1 in /runtimes/huggingf…
dependabot[bot] Oct 16, 2024
6baf1d3
build(deps): bump mlflow from 2.16.2 to 2.17.0 in /runtimes/mlflow (#…
dependabot[bot] Oct 16, 2024
abfd99b
build(deps-dev): bump torchmetrics in /runtimes/mlflow (#1923)
dependabot[bot] Oct 16, 2024
f933d14
build(deps): bump importlib-resources from 6.4.0 to 6.4.5 (#1917)
dependabot[bot] Oct 16, 2024
03ec673
build(deps): bump aiokafka from 0.10.0 to 0.11.0 (#1916)
dependabot[bot] Oct 16, 2024
8ac1d28
build(deps): bump pyspark from 3.5.2 to 3.5.3 in /runtimes/mllib (#1915)
dependabot[bot] Oct 16, 2024
1122340
relax pydantic dependencies (#1928)
lemonhead94 Oct 20, 2024
374cd72
build(deps): Upgrade fastapi and starlette (#1934)
sakoush Oct 21, 2024
64901b2
build(deps-dev): bump torchmetrics in /runtimes/mlflow (#1929)
dependabot[bot] Oct 21, 2024
f8ddfbf
build(deps): bump pillow from 10.4.0 to 11.0.0 in /runtimes/huggingfa…
dependabot[bot] Oct 21, 2024
3ec64dd
build(deps-dev): bump torchmetrics from 1.4.0.post0 to 1.5.0 (#1930)
dependabot[bot] Oct 21, 2024
d160666
build(deps): bump pandas from 2.2.2 to 2.2.3 (#1933)
dependabot[bot] Oct 21, 2024
1ce29e6
Re-generate license info (#1935)
github-actions[bot] Oct 21, 2024
f2b9941
build(deps): bump werkzeug from 3.0.3 to 3.0.6 in /runtimes/alibi-det…
dependabot[bot] Nov 11, 2024
dc45bd9
build(deps-dev): bump werkzeug from 3.0.3 to 3.0.6 (#1940)
dependabot[bot] Nov 11, 2024
0d20436
build(deps): bump werkzeug from 3.0.3 to 3.0.6 in /runtimes/huggingfa…
dependabot[bot] Nov 11, 2024
5fbf96b
build(deps): bump werkzeug from 3.0.3 to 3.0.6 in /runtimes/mlflow (#…
dependabot[bot] Nov 11, 2024
d68cc42
build(deps-dev): bump torchmetrics in /runtimes/mlflow (#1952)
dependabot[bot] Nov 11, 2024
e7ca25a
build(deps): bump optimum from 1.23.1 to 1.23.3 in /runtimes/huggingf…
dependabot[bot] Nov 11, 2024
ed53c7f
build(deps-dev): bump waitress from 3.0.0 to 3.0.1 (#1947)
dependabot[bot] Nov 11, 2024
e73ee5a
build(deps-dev): bump waitress from 3.0.0 to 3.0.1 in /runtimes/mlflo…
dependabot[bot] Nov 11, 2024
cf7fc38
build(deps): bump xgboost from 2.1.1 to 2.1.2 in /runtimes/xgboost (#…
dependabot[bot] Nov 11, 2024
a25735b
build(deps): bump mlflow from 2.17.0 to 2.17.2 in /runtimes/mlflow (#…
dependabot[bot] Nov 11, 2024
c5cc74a
Update FastAPI version constraint (#1954)
sergioave Nov 12, 2024
7063618
Forward rest parameters to model (#1921)
idlefella Nov 19, 2024
3f2c043
build(deps): bump mlflow from 2.17.2 to 2.18.0 in /runtimes/mlflow (#…
dependabot[bot] Nov 26, 2024
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
42 changes: 42 additions & 0 deletions CHANGELOG.md
Original file line number Diff line number Diff line change
@@ -1,6 +1,47 @@
# Changelog


<a name="1.6.1"></a>
## [1.6.1](https://github.com/SeldonIO/MLServer/releases/tag/1.6.1) - 10 Sep 2024

<!-- Release notes generated using configuration in .github/release.yml at 1.6.1 -->

## Overview

### Features
MLServer now offers an option to use pre-existing Python environments by specifying a path to the environment to be used - by [@idlefella](https://github.com/idlefella) in ([#1891](https://github.com/SeldonIO/MLServer/issues/1891))

### Releases
MLServer released catboost runtime which allows serving [catboost](https://catboost.ai/) models with MLServer - by [@sakoush](https://github.com/sakoush) in ([#1839](https://github.com/SeldonIO/MLServer/issues/1839))

### Fixes
* Kafka json byte encoding fix to match rest server by [@DerTiedemann](https://github.com/DerTiedemann) and [@sakoush](https://github.com/sakoush) in ([#1622](https://github.com/SeldonIO/MLServer/issues/1622))
* Prometheus interceptor fix for gRPC streaming by [@RobertSamoilescu](https://github.com/RobertSamoilescu) in ([#1858](https://github.com/SeldonIO/MLServer/issues/1858))


## What's Changed
* Re-generate License Info by [@github-actions](https://github.com/github-actions) in https://github.com/SeldonIO/MLServer/pull/1812
* Update CHANGELOG by [@github-actions](https://github.com/github-actions) in https://github.com/SeldonIO/MLServer/pull/1830
* Update release.yml to include catboost by [@sakoush](https://github.com/sakoush) in https://github.com/SeldonIO/MLServer/pull/1839
* Fix kafka json byte encoding to match rest server by [@DerTiedemann](https://github.com/DerTiedemann) in https://github.com/SeldonIO/MLServer/pull/1622
* Included Prometheus interceptor support for gRPC streaming by [@RobertSamoilescu](https://github.com/RobertSamoilescu) in https://github.com/SeldonIO/MLServer/pull/1858
* Run gRPC test serially by [@RobertSamoilescu](https://github.com/RobertSamoilescu) in https://github.com/SeldonIO/MLServer/pull/1872
* Re-generate License Info by [@github-actions](https://github.com/github-actions) in https://github.com/SeldonIO/MLServer/pull/1886
* Feature/support existing environments by [@idlefella](https://github.com/idlefella) in https://github.com/SeldonIO/MLServer/pull/1891
* Fix tensorflow upperbound macos by [@RobertSamoilescu](https://github.com/RobertSamoilescu) in https://github.com/SeldonIO/MLServer/pull/1901
* ci: Merge change for release 1.6.1 by [@RobertSamoilescu](https://github.com/RobertSamoilescu) in https://github.com/SeldonIO/MLServer/pull/1902
* Bump preflight to 1.10.0 by [@RobertSamoilescu](https://github.com/RobertSamoilescu) in https://github.com/SeldonIO/MLServer/pull/1903
* ci: Merge change for release 1.6.1 [2] by [@RobertSamoilescu](https://github.com/RobertSamoilescu) in https://github.com/SeldonIO/MLServer/pull/1904

## New Contributors
* [@DerTiedemann](https://github.com/DerTiedemann) made their first contribution in https://github.com/SeldonIO/MLServer/pull/1622
* [@idlefella](https://github.com/idlefella) made their first contribution in https://github.com/SeldonIO/MLServer/pull/1891

**Full Changelog**: https://github.com/SeldonIO/MLServer/compare/1.6.0...1.6.1

[Changes][1.6.1]


<a name="1.6.0"></a>
## [1.6.0](https://github.com/SeldonIO/MLServer/releases/tag/1.6.0) - 26 Jun 2024

Expand Down Expand Up @@ -509,6 +550,7 @@ To learn more about how to use MLServer directly from the MLflow CLI, check out
[Changes][1.1.0]


[1.6.1]: https://github.com/SeldonIO/MLServer/compare/1.6.0...1.6.1
[1.6.0]: https://github.com/SeldonIO/MLServer/compare/1.5.0...1.6.0
[1.5.0]: https://github.com/SeldonIO/MLServer/compare/1.4.0...1.5.0
[1.4.0]: https://github.com/SeldonIO/MLServer/compare/1.3.5...1.4.0
Expand Down
Binary file added docs-gb/.gitbook/assets/kserve-logo.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Binary file added docs-gb/.gitbook/assets/seldon-core-logo.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
26 changes: 26 additions & 0 deletions docs-gb/PULL_REQUEST_TEMPLATE/pull_request_template.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,26 @@
# Pull Request

## Description
<!-- Provide a brief description of the purpose of this pull request -->

## Changes Made
<!-- Describe the changes introduced by this pull request -->

## Related Issues
<!-- Mention any related issues or tickets that are being addressed by this pull request -->

## Screenshots (if applicable)
<!-- Add screenshots or images to visually represent the changes, if applicable -->

## Checklist
<!-- Make sure to check the items below before submitting your pull request -->

- [ ] Code follows the project's style guidelines
- [ ] All tests related to the changes pass successfully
- [ ] Documentation is updated (if necessary)
- [ ] Code is reviewed by at least one other team member
- [ ] Any breaking changes are communicated and documented

## Additional Notes
<!-- Add any additional notes or context that may be helpful for reviewers -->

151 changes: 151 additions & 0 deletions docs-gb/README.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,151 @@
# MLServer

An open source inference server for your machine learning models.

[![video_play_icon](https://user-images.githubusercontent.com/10466106/151803854-75d17c32-541c-4eee-b589-d45b07ea486d.png)](https://www.youtube.com/watch?v=aZHe3z-8C_w)

## Overview

MLServer aims to provide an easy way to start serving your machine learning
models through a REST and gRPC interface, fully compliant with [KFServing's V2
Dataplane](https://docs.seldon.io/projects/seldon-core/en/latest/reference/apis/v2-protocol.html)
spec. Watch a quick video introducing the project [here](https://www.youtube.com/watch?v=aZHe3z-8C_w).

- Multi-model serving, letting users run multiple models within the same
process.
- Ability to run [inference in parallel for vertical
scaling](https://mlserver.readthedocs.io/en/latest/user-guide/parallel-inference.html)
across multiple models through a pool of inference workers.
- Support for [adaptive
batching](https://mlserver.readthedocs.io/en/latest/user-guide/adaptive-batching.html),
to group inference requests together on the fly.
- Scalability with deployment in Kubernetes native frameworks, including
[Seldon Core](https://docs.seldon.io/projects/seldon-core/en/latest/graph/protocols.html#v2-kfserving-protocol) and
[KServe (formerly known as KFServing)](https://kserve.github.io/website/modelserving/v1beta1/sklearn/v2/), where
MLServer is the core Python inference server used to serve machine learning
models.
- Support for the standard [V2 Inference Protocol](https://docs.seldon.io/projects/seldon-core/en/latest/reference/apis/v2-protocol.html) on
both the gRPC and REST flavours, which has been standardised and adopted by
various model serving frameworks.

You can read more about the goals of this project on the [initial design
document](https://docs.google.com/document/d/1C2uf4SaAtwLTlBCciOhvdiKQ2Eay4U72VxAD4bXe7iU/edit?usp=sharing).

## Usage

You can install the `mlserver` package running:

```bash
pip install mlserver
```

Note that to use any of the optional [inference runtimes](#inference-runtimes),
you'll need to install the relevant package.
For example, to serve a `scikit-learn` model, you would need to install the
`mlserver-sklearn` package:

```bash
pip install mlserver-sklearn
```

For further information on how to use MLServer, you can check any of the
[available examples](#examples).

## Inference Runtimes

Inference runtimes allow you to define how your model should be used within
MLServer.
You can think of them as the **backend glue** between MLServer and your machine
learning framework of choice.
You can read more about [inference runtimes in their documentation
page](./docs/runtimes/index.md).

Out of the box, MLServer comes with a set of pre-packaged runtimes which let
you interact with a subset of common frameworks.
This allows you to start serving models saved in these frameworks straight
away.
However, it's also possible to **[write custom
runtimes](./docs/runtimes/custom.md)**.

Out of the box, MLServer provides support for:

| Framework | Supported | Documentation |
| ------------- | --------- | ---------------------------------------------------------------- |
| Scikit-Learn | ✅ | [MLServer SKLearn](./runtimes/sklearn) |
| XGBoost | ✅ | [MLServer XGBoost](./runtimes/xgboost) |
| Spark MLlib | ✅ | [MLServer MLlib](./runtimes/mllib) |
| LightGBM | ✅ | [MLServer LightGBM](./runtimes/lightgbm) |
| CatBoost | ✅ | [MLServer CatBoost](./runtimes/catboost) |
| Tempo | ✅ | [`github.com/SeldonIO/tempo`](https://github.com/SeldonIO/tempo) |
| MLflow | ✅ | [MLServer MLflow](./runtimes/mlflow) |
| Alibi-Detect | ✅ | [MLServer Alibi Detect](./runtimes/alibi-detect) |
| Alibi-Explain | ✅ | [MLServer Alibi Explain](./runtimes/alibi-explain) |
| HuggingFace | ✅ | [MLServer HuggingFace](./runtimes/huggingface) |

## Supported Python Versions

🔴 Unsupported

🟠 Deprecated: To be removed in a future version

🟢 Supported

🔵 Untested

| Python Version | Status |
| -------------- | ------ |
| 3.7 | 🔴 |
| 3.8 | 🔴 |
| 3.9 | 🟢 |
| 3.10 | 🟢 |
| 3.11 | 🔵 |
| 3.12 | 🔵 |

## Examples

To see MLServer in action, check out [our full list of
examples](./docs/examples/index.md).
You can find below a few selected examples showcasing how you can leverage
MLServer to start serving your machine learning models.

- [Serving a `scikit-learn` model](./docs/examples/sklearn/README.md)
- [Serving a `xgboost` model](./docs/examples/xgboost/README.md)
- [Serving a `lightgbm` model](./docs/examples/lightgbm/README.md)
- [Serving a `catboost` model](./docs/examples/catboost/README.md)
- [Serving a `tempo` pipeline](./docs/examples/tempo/README.md)
- [Serving a custom model](./docs/examples/custom/README.md)
- [Serving an `alibi-detect` model](./docs/examples/alibi-detect/README.md)
- [Serving a `HuggingFace` model](./docs/examples/huggingface/README.md)
- [Multi-Model Serving with multiple frameworks](./docs/examples/mms/README.md)
- [Loading / unloading models from a model repository](./docs/examples/model-repository/README.md)

## Developer Guide

### Versioning

Both the main `mlserver` package and the [inference runtimes
packages](./docs/runtimes/index.md) try to follow the same versioning schema.
To bump the version across all of them, you can use the
[`./hack/update-version.sh`](./hack/update-version.sh) script.

We generally keep the version as a placeholder for an upcoming version.

For example:

```bash
./hack/update-version.sh 0.2.0.dev1
```

### Testing

To run all of the tests for MLServer and the runtimes, use:

```bash
make test
```

To run run tests for a single file, use something like:

```bash
tox -e py3 -- tests/batch_processing/test_rest.py
```
50 changes: 50 additions & 0 deletions docs-gb/SUMMARY.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,50 @@
* [MLServer](README.md)
* [Getting Started](getting-started.md)
* [User Guide](user-guide/index.md)
* [Content Types (and Codecs)](user-guide/content-type.md)
* [OpenAPI Support](user-guide/openapi.md)
* [Parallel Inference](user-guide/parallel-inference.md)
* [Adaptive Batching](user-guide/adaptive-batching.md)
* [Custom Inference Runtimes](user-guide/custom.md)
* [Metrics](user-guide/metrics.md)
* [Deployment](user-guide/deployment/README.md)
* [Seldon Core](user-guide/deployment/seldon-core.md)
* [KServe](user-guide/deployment/kserve.md)
* [Streaming](user-guide/streaming.md)
* [Inference Runtimes](runtimes/README.md)
* [SKLearn](runtimes/sklearn.md)
* [XGBoost](runtimes/xgboost.md)
* [MLFlow](runtimes/mlflow.md)
* [Spark MlLib](runtimes/mllib.md)
* [LightGBM](runtimes/lightgbm.md)
* [Catboost](runtimes/catboost.md)
* [Alibi-Detect](runtimes/alibi-detect.md)
* [Alibi-Explain](runtimes/alibi-explain.md)
* [HuggingFace](runtimes/huggingface.md)
* [Custom](runtimes/custom.md)
* [Reference](reference/README.md)
* [MLServer Settings](reference/settings.md)
* [Model Settings](reference/model-settings.md)
* [MLServer CLI](reference/cli.md)
* [Python API](reference/python-api/README.md)
* [MLModel](reference/api/model.md)
* [Types](reference/api/types.md)
* [Codecs](reference/api/codecs.md)
* [Metrics](reference/api/metrics.md)
* [Examples](examples/README.md)
* [Serving Scikit-Learn models](examples/sklearn/README.md)
* [Serving XGBoost models](examples/xgboost/README.md)
* [Serving LightGBM models](examples/lightgbm/README.md)
* [Serving MLflow models](examples/mlflow/README.md)
* [Serving a custom model](examples/custom/README.md)
* [Serving Alibi-Detect models](examples/alibi-detect/README.md)
* [Serving HuggingFace Transformer Models](examples/huggingface/README.md)
* [Multi-Model Serving](examples/mms/README.md)
* [Model Repository API](examples/model-repository/README.md)
* [Content Type Decoding](examples/content-type/README.md)
* [Custom Conda environments in MLServer](examples/conda/README.md)
* [Serving a custom model with JSON serialization](examples/custom-json/README.md)
* [Serving models through Kafka](examples/kafka/README.md)
* [Streaming](examples/streaming/README.md)
* [Deploying a Custom Tensorflow Model with MLServer and Seldon Core](examples/cassava/README.md)
* [Changelog](changelog.md)
Loading
Loading