Ask or search…


Integrating Arize with model serving and tooling platform, Spell
Arize helps you visualize your model performance, understand drift & data quality issues, and share insights learned from your models. Spell is an end-to-end ML platform that provides infrastructure for company to deploy and train models.
Read more about the platforms on our partnership announcement.
Read more about the platforms on our partnership announcement
You can either work through on Colab, or follow the steps below for your own model!
Terminal Only
Colab Notebook
Step 1: Logging into spell via command line.
$ spell login
Step 2: Train and create model with spell.
$ spell run \
--github-url \
--machine-type cpu \
--mount public/tutorial/churn_data/:/mnt/churn_prediction/ \
--pip arize --pip lightgbm \
-- python arize/
Step 3: Add your Arize API_KEY and SPACE_KEY to and You can find your Arize credential details here​
Step 4: Creating your model your model and serving it.
$ spell model create churn-prediction 'runs/$RUN_ID'
$ spell server serve \
--node-group default \
--min-pods 1 --max-pods 3 \
--target-requests-per-second 100 \
--pip lightgbm --pip arize \
churn-prediction:v1 # or
Step 5: Test your working instance, send in some data, and see that your model is observable on Arize.
$ curl -X POST -d '@test_payload.txt' \
Spell Integration Tutorial
​Colab Link​