A recent study alleges that LM Arena—the team behind the popular Chatbot Arena—has manipulated its benchmark evaluations to favor select AI labs. Critics argue this approach undermines the fairness of widely recognized scoring methods, fueling demands for greater transparency and accountability in the process.
Back to Top / Wednesday, April 30, 2025, 8:20 pm / permalink 3792 / 6 stories in 10 months
OpenAI Investigates Deceptive Behavior in Chatbot Models / 5 months
ChatGPT safety update introduces parental controls and age prediction / 5 months
OpenAI rethinks GPT‑5 after users cry foul / 6 months
GPT-5 update sparks backlash and swift fixes by OpenAI / 6 months
ChatGPT Agent Launch Sparks High Demand Amid Cautionary Warnings / 7 months
Grok Chatbot’s Antisemitic Rants Stir Outrage Amid Shocking Hitler Praise / 7 months
ChatGPT O3 Model Sabotages Shutdown Command in Test / 9 months
NorthFeed Inc.
Disclaimer: The information provided on this website is intended for general informational purposes only. While we strive for accuracy, we do not guarantee the completeness or reliability of the content. Users are encouraged to verify all details independently. We accept no liability for errors, omissions, or any decisions made based on this information.