LLM Benchmark Table

Toggle Dark Mode
Model TOTAL Pass Refine Fail Refusal $ mToK Reason STEM Utility Code Censor
Model 1 1000 800 150 30 20 $5m High Accuracy Yes High Yes No
What is the LLM Benchmark Table?
The LLM Benchmark Table provides a comparative overview of various AI models based on performance metrics.
How are the models evaluated?
Models are evaluated based on a series of tests that measure their accuracy, efficiency, and adaptability in different scenarios.