Description
Deciding whose values an LLM should align with is a fundamental problem. Current frameworks (like helpfulness, harmlessness, honesty) are themselves value-laden and can conflict. There's a risk of a small group of developers imposing their values on a global user base, especially since decisions about values are often made implicitly.