Added spaCy support. You can now use spaCy models as pipeline components.
Added transformers support. You can now use HuggingFace Transformers models as pipeline components. At this time, we only support specific models. Consult the docs for more information.
Fixed a critical bug causing ONNX models to not run on GPU. You should upgrade to this version if you are using ONNX models that require GPU computation.
onnxruntime is now an optional dependency. While you can still deploy ONNX models without the runtime, you will need to install the extra with pip install --upgrade cerebrium[onnxruntime] to test you Conduit locally.
Added support for onnxruntime-gpu as an optional dependency. You can install the extra with pip install --upgrade cerebrium[onnxruntime-gpu].
New updates to the client post 0.4.1 are now backwards compatible and now will not potentially break existing deployments. We will now support every version from 0.4.1 going forward for the foreseeable future and will not be supporting older versions of the client.
Please upgrade to the latest version of the client to ensure you are using the latest features and fixes. If you want to use a new version of the client in your deployment, simply upgrade and redeploy!
Added the ability to define a Conduit object directly.
We have added external monitoring tool support! You can now monitor your deployments with Censius by adding a logger to the Conduit object.
While it is available to use for all users, this functionality is currently in active development and may change in the coming weeks based
on feedback. We are also working on adding support for other monitoring tools, with Arize being on track to be released in Q1 2023.
Fixed a bug where the client was unable to deploy scikit-learn models, XGBoost Regressor models, and ONNX models.
Reworked the response signature for a model. Now returns a JSON object with 3 fields:
result: The data returned by the deployed Conduit.
run_id: The ID of the run.
prediction_ids: The prediction IDs of each prediction made. Used to track/update the predictions logged to monitoring tools.
Added pre-built deployment for the following LLMs:
We made large changes to our infrastructure to be more reliable and faster. This means faster deployment, less downtime and roughly half the inference time!
Support for ONNX models with Python 3.7-3.9 (can be used as a single model or
as the initial model of a flow). We are working on supporting Python 3.10
and 3.11!