Similar Posts
But what is a neural network? | Deep learning chapter 1
Byn0cadminWhat are the neurons, why are there layers, and what is the math underlying it? Typo correction: At 14 minutes 45 seconds, the last index on the bias vector is n, when it’s supposed to in fact be a k. Thanks for the sharp eyes that caught that! There are two neat things about this…
What does it mean for computers to understand language? | LM1
Byn0cadminAn introduction to language modeling, followed by an explanation of the N-Gram language model! Sources (includes the entire series): https://docs.google.com/document/d/1e… Chapters0:00 Introduction1:39 What is NLP?2:45 What is a Language Model?4:38 N-Gram Language Model7:20 Inference9:18 Outro
The Attention Mechanism in Large Language Models
Byn0cadminAttention mechanisms are crucial to the huge boom LLMs have recently had.In this video you’ll see a friendly pictorial explanation of how attention mechanisms work in Large Language Models.This is the first of a series of three videos on Transformer models. https://www.youtube.com/watch?v=OxCpWwDCDFQ
Fine-tuning Large Language Models (LLMs) | w/ Example Code
Byn0cadminThis is the 5th video in a series on using large language models (LLMs) in practice. Here, I discuss how to fine-tune an existing LLM for a particular use case and walk through a concrete example with Python code.
The math behind Attention: Keys, Queries, and Values matrices
Byn0cadminThis is the second of a series of 3 videos where we demystify Transformer models and explain them with visuals and friendly examples. 00:00 Introduction01:18 Recap: Embeddings and Context04:46 Similarity11:09 Attention20:46 The Keys and Queries Matrices25:02 The Values Matrix28:41 Self and Multi-head attention33:54: Conclusion
Practical LLM Security: Takeaways From a Year in the Trenches
Byn0cadminOct 9, 2024As LLMs are being integrated into more and more applications, security standards for these integrations have lagged behind. Most security research either focuses 1) on social harms, biases exhibited by LLMs, and other content moderation tasks, or 2) zooms in on the LLM itself and ignores the applications that are built around them….
