Selene offers a simple and flexible option to run your evaluations using a python SDK.

You can run an evaluation using Atla default metrics or a custom metric you created using the Eval Copilot.

Selene not only outputs a score, but also a helpful Chain-of-Thought critique. You can use these to generate valuable insights across your LLM lifecycle.

Next steps:


Read about how you can use Selene’s scores and critiques to improve your LLM outputs here.