BigCodeBench: Benchmarking Large Language Models on Solving Practical and Challenging Programming Tasks Jun 18, 2024 • 43
Falcon 2: An 11B parameter pretrained language model and VLM, trained on over 5000B tokens tokens and 11 languages May 24, 2024 • 25
CyberSecEval 2 - A Comprehensive Evaluation Framework for Cybersecurity Risks and Capabilities of Large Language Models May 24, 2024 • 21
The Open Medical-LLM Leaderboard: Benchmarking Large Language Models in Healthcare Apr 19, 2024 • 126
Introducing the LiveCodeBench Leaderboard - Holistic and Contamination-Free Evaluation of Code LLMs Apr 16, 2024 • 14
Introducing ConTextual: How well can your Multimodal model jointly reason over text and image in text-rich scenes? Mar 5, 2024 • 4
Introducing the Open Ko-LLM Leaderboard: Leading the Korean LLM Evaluation Ecosystem Feb 20, 2024 • 3
NPHardEval Leaderboard: Unveiling the Reasoning Abilities of Large Language Models through Complexity Classes and Dynamic Updates Feb 2, 2024 • 3
Introducing the Enterprise Scenarios Leaderboard: a Leaderboard for Real World Use Cases Jan 31, 2024 • 3
The Hallucinations Leaderboard, an Open Effort to Measure Hallucinations in Large Language Models Jan 29, 2024 • 17
A guide to setting up your own Hugging Face leaderboard: an end-to-end example with Vectara's hallucination leaderboard Jan 12, 2024 • 6
Running 497 😻 Open Source Ai Year In Review 2024 What happened in open-source AI this year, and what’s next?