Add to Favourites
To login click here

Patronus AI recently launched the first automated evaluation and security platform that helps companies safely use large language models (LLMs). The platform enables enterprise development teams to score model performance, generate adversarial test cases, benchmark models, and more. It automates and scales the manual and costly model evaluation methods prevalent in the enterprise today, enabling organizations to confidently deploy LLMs while minimizing the risk of model failures and misaligned outputs. Patronus AI was founded by machine learning experts Anand Kannappan and Rebecca Qian, and is backed by a $3 million seed funding round led by Lightspeed Venture Partners. Key features of the platform include scoring, test generation, and benchmarking.