MIT News • 2/19/2026

MIT has developed a new method aimed at exposing biases, moods, personalities, and abstract concepts hidden in large language models (LLMs). This method could help root out vulnerabilities and improve the safety and performance of LLMs.
Advertisement
















