💭 if we could somehow totally delete an llms understanding of a given concept (sounds basically impossible) then we could use teaching them to understand that concept as a benchmark for teaching ability. Could also just make up new things to teach but it's not grounded and would leak into training data