Large Language Models explained briefly
Timestamps:
0:00 – Who this was made for
0:41 – What are large language models?
7:48 – Where to learn more
Timestamps:
0:00 – Who this was made for
0:41 – What are large language models?
7:48 – Where to learn more
The attention mechanism is well known for its use in Transformers. But where does it come from? It’s origins lie in fixing a strange problems of RNNs. Chapters0:00 Introduction0:22 Machine Translation2:01 Attention Mechanism8:04 Outro
What is Generative AI and how does it work? What are common applications for Generative AI? Watch this video to learn all about Generative AI, including common applications, model types, and the fundamentals for how to use it.
This is the 5th video in a series on using large language models (LLMs) in practice. Here, I discuss how to fine-tune an existing LLM for a particular use case and walk through a concrete example with Python code.
For more information about Stanford’s Artificial Intelligence programs visit: https://stanford.io/ai https://www.youtube.com/watch?v=Bl4Feh_Mjvo To follow along with the course, visit:https://cs229.stanford.edu/syllabus-s… Tengyu MaAssistant Professor of Computer Sciencehttps://ai.stanford.edu/~tengyuma/ Christopher RéAssociate Professor of Computer Sciencehttps://cs.stanford.edu/~chrismre/
To learn more, I highly recommend the book by Michael Nielsenhttp://neuralnetworksanddeeplearning….The book walks through the code behind the example in these videos, which you can find here:https://github.com/mnielsen/neural-ne… MNIST database:http://yann.lecun.com/exdb/mnist/ Also check out Chris Olah’s blog:http://colah.github.io/His post on Neural networks and topology is particular beautiful, but honestly all of the stuff there is great. And if…
To follow along with the course, visit the course website:https://deepgenerativemodels.github.io/ Stefano ErmonAssociate Professor of Computer Science, Stanford Universityhttps://cs.stanford.edu/~ermon/ https://www.youtube.com/watch?v=XZ0PMRWXBEU