Large Language Models
Anthropic launched a programme to fund third-party AI evaluations, focusing on safety and advanced capabilities. The initiative covers areas like cybersecurity, multilingual skills, and societal impacts, aiming to improve AI safety across the industry.
CriticGPT, a GPT-4 model, identifies ChatGPT code errors. Human reviewers using it outperform others by 60%. It shows 63% preference in catching real bugs.
OpenAI's new content moderation system uses advanced methods to detect undesired content online. It aims to identify various types of problematic material more effectively than current models.
On June 5, 2024, the Mistral AI team announced three different entry points for specialising their AI models:
Researchers at OpenAI have developed new methods for interpreting the neural activity within language models, providing a glimpse into the concepts and
Reddit's deals with AI companies for training data highlight the growing value of user-generated content and raise questions about data ownership.