Description
Ensuring one LLM agent is safe doesn't guarantee safety when multiple LLM agents interact. Interactions can lead to suboptimal outcomes for everyone, and groups of LLM agents might develop unexpected collective behaviors or even collude in undesirable ways. Because many LLMs share similar foundations, they might also be prone to correlated failures (the same problem affecting many of them simultaneously).