Demystifying attention, the key mechanism inside transformers and LLMs.