💭 I dont think there's any serious sota breaks in recent llm models, at least for my practical usage. They're good and bad at roughly the same things to roughly the same degree and differences I can notice are mostly standard variance I think.