Prompt Testing

You can easily test different prompt and model combinations. Variable parametrization is also supported.

Results will show the LLM response, latency, cost and optionally the Evaluator scores:

Last updated