This is very interesting. I don't see much discussion of interpretability in day to the day discourse of AI builders. I wonder if everyone assumes it to either be solved, or to be too out of reach to bother stopping and thinking about.
brendanashworth•4m ago
Is there a reason people don't use SHAP [1] to interpret language models more often? The in-context attribution of outputs seems very similar.
pbmango•5m ago