Learn about Selene
Selene is the family of state-of-the-art LLM Judge models built by Atla. Selene is trained specifically to evaluate generative AI responses. It excels at evaluating LLM outputs by other models involving language, coding, math, chat, RAG contexts, and more. You can use Selene to evaluate your LLM outputs - whether you’re building on OpenAI, Anthropic, Mistral, Meta, or your own LLM.
Selene evaluates your outputs against your scoring criteria and generates a score and a Chain-of-Thought critique. We have built Selene with scale and flexibility in mind. You can use the score and critique generated for a variety of use cases:
Learn about Selene
Selene is the family of state-of-the-art LLM Judge models built by Atla. Selene is trained specifically to evaluate generative AI responses. It excels at evaluating LLM outputs by other models involving language, coding, math, chat, RAG contexts, and more. You can use Selene to evaluate your LLM outputs - whether you’re building on OpenAI, Anthropic, Mistral, Meta, or your own LLM.
Selene evaluates your outputs against your scoring criteria and generates a score and a Chain-of-Thought critique. We have built Selene with scale and flexibility in mind. You can use the score and critique generated for a variety of use cases: