>Transformers are a machine-learning model at the foundation of many state-of-the-art systems in modern AI, originally proposed in [arXiv:1706.03762]. In this post, we are going to build a generalization of Transformer models that can operate on (almost) arbitrary structures such as functions, graphs, probability distributions, not just matrices and vectors.
>[...]
>This work is part of a series of similar ideas exploring machine learning through abstract diagrammatical means.
https://cybercat.institute/2025/02/12/transformers-applicati...
pawel ... ?
But then ai realized I would always forget the names for all the mathy c' words - commutativity commutativity, qssociativity... and for the first time I could actually remember commutativity and what it means, just because he tied it into a graphical representation (which actually made me laugh out loud because, initially, I thought it was a joke). So the concept of "x + y = y + x" always made sense to me but never really stuck like the graphical representation, which also made me remember its name for the first time.
I am sold.
Aka one of my favorite axioms: https://www.penny-arcade.com/comic/2004/03/19/green-blackboa...
From the perspective of the lambda calculus for example, the duplication of the addition node in "When Adding met Copying" [2] mirrors exactly the iterative duplication of lambda terms - ie. something like (λx.x x) M!
[1]: https://ezb.io/thoughts/interaction_nets/lambda_calculus/202...
[2]: https://graphicallinearalgebra.net/2015/05/12/when-adding-me...
lorenzo_medici•6h ago