Data Imports
Overview
The imports command group manages data import jobs that bring data from external sources into Arize. There are two sub-groups:
imports files— Import from cloud object storage (S3, GCS, Azure Blob Storage)imports tables— Import from data warehouses (BigQuery, Snowflake, Databricks)
File Import Commands
| Command | Description | Client Method |
|---|---|---|
imports files list |
List file import jobs | get_all_file_import_jobs |
imports files get |
Get a file import job by ID | get_file_import_job |
imports files create |
Create a file import job | create_file_import_job |
imports files delete |
Delete a file import job | delete_file_import_job |
Table Import Commands
| Command | Description | Client Method |
|---|---|---|
imports tables list |
List table import jobs | get_all_table_import_jobs |
imports tables get |
Get a table import job by ID | get_table_import_job |
imports tables create |
Create a table import job | create_table_import_job |
imports tables delete |
Delete a table import job | delete_table_import_job |
File Imports
imports files list
arize_toolkit imports files list
Lists all file import jobs in the current space.
Example
$ arize_toolkit imports files list
File Import Jobs
┌──────────┬──────────┬────────────┬────────────┐
│ id │ jobId │ jobStatus │ createdAt │
├──────────┼──────────┼────────────┼────────────┤
│ f1 │ j-abc │ completed │ 2025-01-10 │
│ f2 │ j-def │ running │ 2025-02-01 │
└──────────┴──────────┴────────────┴────────────┘
imports files get
arize_toolkit imports files get JOB_ID
Retrieves details for a file import job, including file counts and schema.
Arguments
JOB_ID— The import job ID.
Example
arize_toolkit --json imports files get "j-abc123"
imports files create
arize_toolkit imports files create --blob-store STORE --bucket BUCKET --prefix PREFIX --model MODEL --model-type TYPE --schema JSON [OPTIONS]
Creates a new file import job from cloud object storage.
Required Options
--blob-store— Cloud provider:s3,gcs, orazure.--bucket— Bucket or container name.--prefix— Object prefix/path within the bucket.--model— Target model name.--model-type— Model type:classification,regression,ranking,object_detection,multi-class, orgenerative.--schema— Model schema as a JSON string or@filepath.
Optional Options
--model-version— Model version string.--environment— Environment:production,validation,training,tracing. Defaults toproduction.--dry-run— Validate without importing.--batch-id— Batch ID for validation data.--azure-tenant-id— Azure tenant ID (Azure only).--azure-storage-account— Azure storage account name (Azure only).
Schema Format
The schema defines column mappings. Pass as inline JSON or from a file:
{
"predictionId": "id_column",
"timestamp": "ts_column",
"predictionLabel": "pred_column",
"actualLabel": "actual_column",
"features": "feature_"
}
Example
arize_toolkit imports files create \
--blob-store s3 \
--bucket "my-data-bucket" \
--prefix "models/fraud/2025-01/" \
--model "fraud-detection-v3" \
--model-type classification \
--schema @schema.json \
--environment production
imports files delete
arize_toolkit imports files delete JOB_ID [--yes]
Deletes a file import job. Prompts for confirmation unless --yes is passed.
Arguments
JOB_ID— The import job ID.
Options
--yes— Skip confirmation.
Example
arize_toolkit imports files delete "j-abc123" --yes
Table Imports
imports tables list
arize_toolkit imports tables list
Lists all table import jobs in the current space.
Example
arize_toolkit imports tables list
imports tables get
arize_toolkit imports tables get JOB_ID
Retrieves details for a table import job.
Arguments
JOB_ID— The import job ID.
Example
arize_toolkit --json imports tables get "j-xyz789"
imports tables create
arize_toolkit imports tables create --table-store STORE --model MODEL --model-type TYPE --schema JSON --table-config JSON [OPTIONS]
Creates a new table import job from a data warehouse.
Required Options
--table-store— Provider:BigQuery,Snowflake, orDatabricks.--model— Target model name.--model-type— Model type:classification,regression,ranking,object_detection,multi-class, orgenerative.--schema— Model schema as a JSON string or@filepath.--table-config— Table configuration as a JSON string or@filepath.
Optional Options
--model-version— Model version string.--environment— Environment:production,validation,training,tracing. Defaults toproduction.--dry-run— Validate without importing.--batch-id— Batch ID for validation data.
Table Configuration Formats
{
"projectId": "my-gcp-project",
"dataset": "ml_data",
"tableName": "predictions"
}
{
"accountID": "abc123",
"schema": "PUBLIC",
"database": "ML_DB",
"tableName": "PREDICTIONS"
}
{
"hostName": "my-workspace.databricks.com",
"endpoint": "/sql/1.0/warehouses/abc",
"port": "443",
"catalog": "main",
"databricksSchema": "default",
"tableName": "predictions"
}
Example
arize_toolkit imports tables create \
--table-store BigQuery \
--model "fraud-detection-v3" \
--model-type classification \
--schema @schema.json \
--table-config '{"projectId":"my-project","dataset":"ml","tableName":"preds"}'
imports tables delete
arize_toolkit imports tables delete JOB_ID [--yes]
Deletes a table import job. Prompts for confirmation unless --yes is passed.
Arguments
JOB_ID— The import job ID.
Options
--yes— Skip confirmation.
Example
arize_toolkit imports tables delete "j-xyz789" --yes