Transformers are described in the document "attention is all you need" and are the architecture used by AI large language models (chatgpt, etc): FUNCTION Main() LOCAL aEmbeddings, aWq, aWk, aWv, aBq, aBk, aBv LOCAL aQ, aK, aV LOCAL aAttentionScores, ...