This data set helps researchers spot harmful stereotypes in LLMs
AI models are riddled with culturally specific biases. A new data set, called SHADES, is designed to help developers combat the problem by spotting harmful stereotypes and other kinds of...