In the news
Study: Some language reward models exhibit political bias
Research from the MIT Center for Constructive Communication finds this effect occurs even when reward models are trained on factual data.
Large language models (LLMs) that drive generative artificial intelligence apps, such as ChatGPT, have been proliferating at lightning speed and have improved to the point that it is often impossible to distinguish between something written through generative AI and human-composed text. However, these models can also sometimes generate false statements or display a political bias.
News