Large Language Models explained briefly
Timestamps:
0:00 – Who this was made for
0:41 – What are large language models?
7:48 – Where to learn more
Timestamps:
0:00 – Who this was made for
0:41 – What are large language models?
7:48 – Where to learn more
Dec 3, 2024One-day workshop on topics in Generative AI IISc-IBM AI Day is being jointly organized by the Centre for Networked Intelligence (with support from Cisco CSR) and IBM-IISc Hybrid Cloud Lab, in collaboration with IBM India Research Lab. The goal of this workshop would be to apprise the audience of Generative AI, a set…
Chapters0:00 Introduction1:54 Neural N-Gram Models6:03 Recurrent Neural Networks11:47 LSTM Cells12:22 Outro
Topics: Overview of course, OptimizationPercy Liang, Associate Professor & Dorsa Sadigh, Assistant Professor – Stanford Universityhttp://onlinehub.stanford.edu/ Associate Professor Percy LiangAssociate Professor of Computer Science and Statistics (courtesy) Assistant Professor Dorsa SadighAssistant Professor in the Computer Science Department & Electrical Engineering Department To follow along with the course schedule and syllabus, visit:https://stanford-cs221.github.io/autumn2019/#schedule artificialintelligencecourse 0:00 Introduction3:30 Why…
https://www.youtube.com/watch?v=9-Jl0dxWQs8 AI Alignment forum post from the Deepmind researchers referenced at the video’s start:https://www.alignmentforum.org/posts/… Anthropic posts about superposition referenced near the end:https://transformer-circuits.pub/2022…https://transformer-circuits.pub/2023… Some added resources for those interested in learning more about mechanistic interpretability, offered by Neel Nanda Mechanistic interpretability paper reading listhttps://www.alignmentforum.org/posts/… Getting started in mechanistic interpretabilityhttps://www.neelnanda.io/mechanistic-… An interactive demo of sparse autoencoders (made…
This one is a bit more symbol-heavy, and that’s actually the point. The goal here is to represent in somewhat more formal terms the intuition for how backpropagation works in part 3 of the series, hopefully providing some connection between that video and other texts/code that you come across later. For more on backpropagation:http://neuralnetworksanddeeplearning….https://github.com/mnielsen/neural-ne…http://colah.github.io/posts/2015-08-… https://colah.github.io/posts/2015-08-Backprop