标签:LLM evaluation
Algomax
Algomax streamlines LLM & RAG model evaluation and enhances prompt development.
Ottic
Platform to evaluate and test LLM-powered applications for faster and reliable product releases.
Helicone AI
Open-source LLM observability platform for monitoring, debugging, and improving AI apps.
EvalMy.AI
Automated AI answer verification service using C3-score for accuracy and efficiency.
Scale
Scale AI provides high-quality training data and platforms for AI development and evaluation.
Future AGI
AI evaluation and optimization platform for automated quality assessment and performance enhancement.
EvalsOne
A platform for evaluating and optimizing generative AI applications.
Confident AI
All-in-one LLM evaluation platform for testing, benchmarking, and improving LLM application performance.
Maxim AI
End-to-end AI evaluation and observability platform for testing and deploying AI applications.
Airtrain.ai LLM Playground
A platform for exploring, curating, and evaluating unstructured datasets and LLMs.
AutoArena
Open-source tool for automated head-to-head evaluation of GenAI systems using LLM judges.
You Rate AI
A platform for rating AI (LLMs) services based on real-world user experiences.