Large language models (LLMs) such as GPT and Llama are driving exceptional innovations in AI, but research aimed at improving their explainability and reliability is constrained by massive resource requirements for examining and adjusting their behavior.

