2 results for tag "arize-experiment"
Manages Arize experiments (named evaluation runs against dataset versions) via `ax experiments list/get/create/export/delete`, accepting runs files with required `example_id` and `output` columns plus optional `evaluations` and `metadata`. Uses REST by default (500-run cap) and Arrow Flight via `--all` for bulk export.
Skills that guide AI coding agents to add observability, run experiments, and optimize prompts for LLM applications using the Arize platform and the ax CLI, with support for tracing, debugging, and production monitoring workflows.