Arize-ai / phoenix

AI Observability & Evaluation
https://docs.arize.com/phoenix
Other
2.98k stars 212 forks source link

🗺 prompt playground #3435

Open mikeldking opened 2 weeks ago

mikeldking commented 2 weeks ago
heralight commented 4 days ago

Hi!

Enhancement proposal

This feature should be similar to #2462 but with more depth. It would involve a simple button to replicate a query into an edit mode, allowing you to replay it. Additionally, it should offer the possibility to add notes on the result iterations, such as rating the quality, format output, etc., on a scale of 1 to 10.

Goal

The goal is to facilitate quick testing of prompts and inputs, enabling evaluation and visualization of progression.

Thank you,

Alexandre

mikeldking commented 4 days ago

Hi!

Enhancement proposal

This feature should be similar to #2462 but with more depth. It would involve a simple button to replicate a query into an edit mode, allowing you to replay it. Additionally, it should offer the possibility to add notes on the result iterations, such as rating the quality, format output, etc., on a scale of 1 to 10.

Goal

The goal is to facilitate quick testing of prompts and inputs, enabling evaluation and visualization of progression.

Thank you,

Alexandre

@heralight Hey! Thanks for the feedback! We have a ton of features coming out with regards to prompt iteration, notably prompt experiments. Stay tuned. It has evaluations built in

Noted on the replay and the annotations:) will give it some thought. We have a few ideas around replaying data at prompts, but haven't thought about human annotations on different prom versions a ton. Would love to hear more.

heralight commented 3 days ago

Very nice! my ideal workflow, would be:

best,