text
stringlengths 177
219
|
---|
llm_intro_0.wav|This is unlike many other language models that you might be familiar with. For example, if you're using Chats GPT or something like that, the model architecture was never released.| |
llm_intro_1.wav|Because this is a 70 billion parameter model, every one of those parameters is stored as two bytes. And so therefore, the parameters file here is 104 gigabytes.| |
llm_intro_2.wav|And it's two bytes because this is a float 16 number as the data type. Now, in addition to these parameters, that's just like a large list of parameters for that neural network.| |
llm_intro_3.wav|You also need something that runs that neural network. And this piece of code is implemented in our run file. Now, this could be a C file or a Python file or any other programming language, really.| |
llm_intro_4.wav|It can be written in any arbitrary language. But C is sort of like a very simple language, just to give you a sense. And it would only require about 500 lines of C with no other dependencies.| |
llm_intro_5.wav|It was only running a seven billion parameter model. A 70B would be running about 10 times slower, but I wanted to give you an idea of sort of just the text generation and what that looks like.| |
llm_intro_6.wav|So not a lot is necessary to run the model. This is a very small package. But the computational complexity really comes in when we'd like to get those parameters.| |
llm_intro_7.wav|Because whatever's in the run.c file, the neural network architecture and sort of the forward pass of that network, everything is algorithmically understood and open and so on.| |
llm_intro_8.wav|So to obtain the parameters, basically the model training, as we call it, is a lot more involved than model inference, which is the part that I showed you earlier.| |
llm_intro_9.wav|So because LLAMA270B is an open source model, we know quite a bit about how it was trained because Meta released that information in paper. So these are some of the numbers of what's involved.| |
llm_intro_10.wav|You basically take a chunk of the internet that is roughly, you should be thinking, 10 terabytes of text. This typically comes from like a crawl of the internet.| |
llm_intro_11.wav|So just imagine just collecting tons of text from all kinds of different websites and collecting it together. So you take a large chunk of internet, then you procure a GPU cluster.| |
llm_intro_12.wav|So these parameters that I showed you in an earlier slide are best kind of thought of as like a zip file of the internet. And in this case, what would come out are these parameters, 140 gigabytes.| |
llm_intro_13.wav|So you can see that the compression ratio here is roughly like 100x, roughly speaking. But this is not exactly a zip file because a zip file is lossless compression.| |
llm_intro_14.wav|What's happening here is a lossy compression. We're just kind of like getting a kind of a gestalt of the text that we trained on. We don't have an identical copy of it in these parameters.| |
llm_intro_15.wav|So if you want to think about state-of-the-art neural networks, like, say, what you might use in chatGPT, or CLOD, or BARD, or something like that, these numbers are off by a factor of 10 or more.| |
llm_intro_16.wav|OK, so what is this neural network really doing? I mentioned that there are these parameters. This neural network basically is just trying to predict the next word in a sequence.| |
llm_intro_17.wav|You can think about it that way. So you can feed in a sequence of words, for example, cat sat on a. This feeds into a neural net. And these parameters are dispersed throughout this neural network.| |
llm_intro_18.wav|And there's neurons, and they're connected to each other, and they all fire in a certain way. You can think about it that way. And out comes a prediction for what word comes next.| |
llm_intro_19.wav|So for example, in this case, this neural network might predict that in this context of four words, the next word will probably be a mat with, say, 97% probability.| |
llm_intro_20.wav|And you can show mathematically that there's a very close relationship between prediction and compression, which is why I allude to this neural network as training it as a compression of the internet.| |
llm_intro_21.wav|And so think about being the neural network, and you're given some amount of words and trying to predict the next word in a sequence. Well, in this case, I'm highlighting here in red| |
llm_intro_22.wav|some of the words that would contain a lot of information. And so, for example, if your objective is to predict the next word, presumably your parameters have to learn a lot of this knowledge.| |
llm_intro_23.wav|And so, in the task of next word prediction, you're learning a ton about the world, and all this knowledge is being compressed into the weights, the parameters.| |
llm_intro_24.wav|We basically generate what comes next, we sample from the model, so we pick a word, and then we continue feeding it back in and get the next word, and continue feeding that back in.| |
llm_intro_25.wav|So for example, if we just run the neural network, or as we say, perform inference, we would get sort of like web page dreams. You can almost think about it that way, right?| |
llm_intro_26.wav|Because this network was trained on web pages, and then you can sort of like let it loose. So on the left, we have some kind of a Java code dream, it looks like.| |
llm_intro_27.wav|In the middle, we have some kind of what looks like almost like an Amazon product dream. And on the right, we have something that almost looks like a Wikipedia article.| |
llm_intro_28.wav|The model network just knows that what comes after ISBN colon is some kind of a number of roughly this length, and it's got all these digits, and it just like puts it in.| |
llm_intro_29.wav|It just kind of like puts in whatever looks reasonable. So it's parroting the training dataset distribution. On the right, the black nose dace, I looked it up, and it is actually a kind of fish.| |
llm_intro_30.wav|And what's happening here is this text verbatim is not found in the training set documents. But this information, if you actually look it up, is actually roughly correct with respect to this fish.| |
llm_intro_31.wav|But again, it's some kind of a lossy compression of the internet. It kind of remembers the gestalt. It kind of knows the knowledge. And it just kind of like goes.| |
llm_intro_32.wav|But for the most part, this is just kind of like hallucinating or like dreaming internet text from its data distribution. Okay, let's now switch gears to how does this network work?| |
llm_intro_33.wav|Now, what's remarkable about this neural net is we actually understand in full detail the architecture. We know exactly what mathematical operations happen at all the different stages of it.| |
llm_intro_34.wav|The problem is that these 100 billion parameters are dispersed throughout the entire neural network. So basically, these billions of parameters are throughout the neural net.| |
llm_intro_35.wav|And all we know is how to adjust these parameters iteratively to make the network as a whole better at the next word prediction task. So we know how to optimize these parameters.| |
llm_intro_36.wav|So we kind of understand that they build and maintain some kind of a knowledge database, but even this knowledge database is very strange and imperfect and weird.| |
llm_intro_37.wav|It will tell you it's Mary Lee Pfeiffer, which is correct. But if you say, who is merely Pfeiffer's son, it will tell you it doesn't know. So this knowledge is weird and it's kind of one-dimensional.| |
llm_intro_38.wav|And you have to sort of like, this knowledge isn't just like stored and can be accessed in all the different ways. You have to sort of like ask it from a certain direction almost.| |
llm_intro_39.wav|And so that's really weird and strange. And fundamentally, we don't really know because all you can kind of measure is whether it works or not and with what probability.| |
llm_intro_40.wav|But right now we kind of treat them mostly as empirical artifacts. We can give them some inputs and we can measure the outputs. We can basically measure their behavior.| |
llm_intro_41.wav|And so I think this requires basically correspondingly sophisticated evaluations to work with these models because they're mostly empirical. So now let's go to how we actually obtain an assistant.| |
llm_intro_42.wav|And this is where we obtain what we call an assistant model, because we don't actually really just want document generators. That's not very helpful for many tasks.| |
llm_intro_43.wav|And the way you obtain these assistant models is fundamentally through the following process. We basically keep the optimization identical, so the training will be the same.| |
llm_intro_44.wav|It's just a next word prediction task. But we're going to swap out the data set on which we are training. So it used to be that we are trying to train on internet documents.| |
llm_intro_45.wav|We're going to now swap it out for data sets that we collect manually. And the way we collect them is by using lots of people. So typically, a company will hire people.| |
llm_intro_46.wav|So we swap out the dataset now, and we train on these Q&A documents. And this process is called fine-tuning. Once you do this, you obtain what we call an assistant model.| |
llm_intro_47.wav|So this assistant model now subscribes to the form of its new training documents. So for example, if you give it a question like, can you help me with this code?| |
llm_intro_48.wav|In the pre-training stage, you get a ton of text from the internet. You need a cluster of GPUs. So these are special purpose computers for these kinds of parallel processing workloads.| |
llm_intro_49.wav|This is not just things that you can buy and best buy. These are very expensive computers. And then you compress the text into this neural network, into the parameters of it.| |
llm_intro_50.wav|This would only potentially take like one day or something like that instead of a few months or something like that. And you obtain what we call an assistant model.| |
llm_intro_51.wav|Then you run a lot of evaluations, you deploy this, and you monitor, collect misbehaviors. And for every misbehavior, you want to fix it. And you go to step on and repeat.| |
llm_intro_52.wav|So you take that, and you ask a person to fill in the correct response. And so the person overwrites the response with the correct one, and this is then inserted as an example into your training data.| |
llm_intro_53.wav|Because fine-tuning is a lot cheaper, you can do this every week, every day, or so on, and companies often will iterate a lot faster on the fine-tuning stage instead of the pre-training stage.| |
llm_intro_54.wav|One other thing to point out is, for example, I mentioned the Lama 2 series. The Lama 2 series actually, when it was released by Meta, contains both the base models and the assistant models.| |
llm_intro_55.wav|If you give it questions, it will just give you more questions, or it will do something like that, because it's just an internet document sampler. So these are not super helpful.| |
llm_intro_56.wav|Now see how in stage two I'm saying and or comparisons? I would like to briefly double click on that because there's also a stage three of fine-tuning that you can optionally go to or continue to.| |
llm_intro_57.wav|The reason that we do this is that in many cases it is much easier to compare candidate answers than to write an answer yourself if you're a human labeler. So consider the following concrete example.| |
llm_intro_58.wav|Suppose that the question is to write a haiku about paperclips or something like that. From the perspective of a labeler, if I'm asked to write a haiku, that might be a very difficult task, right?| |
llm_intro_59.wav|Well, then as a labeler, you could look at these haikus and actually pick the one that is much better. And so in many cases, it is easier to do the comparison instead of the generation.| |
llm_intro_60.wav|And there's a stage three of fine-tuning that can use these comparisons to further fine-tune the model. And I'm not going to go into the full mathematical detail of this.| |
llm_intro_61.wav|I also wanted to show you very briefly one slide showing some of the labeling instructions that we give to humans. So this is an excerpt from the paper InstructGPT by OpenAI.| |
llm_intro_62.wav|One more thing that I wanted to mention is that I've described the process naively as humans doing all of this manual work, but that's not exactly right, and it's increasingly less correct.| |
llm_intro_63.wav|And so for example, you can get these language models to sample answers, and then people sort of like cherry pick parts of answers to create one sort of single best answer.| |
llm_intro_64.wav|Or you can ask these models to try to check your work, or you can try to ask them to create comparisons, and then you're just kind of like in an oversight role over it.| |
llm_intro_65.wav|Okay, finally, I wanted to show you a leaderboard of the current leading large language models out there. So this, for example, is the Chatbot Arena. It is managed by a team at Berkeley.| |
llm_intro_66.wav|And what they do here is they rank the different language models by their ELO rating. And the way you calculate ELO is very similar to how you would calculate it in chess.| |
llm_intro_67.wav|So different chess players play each other, and depending on the win rates against each other, you can calculate their ELO scores. You can do the exact same thing with language models.| |
llm_intro_68.wav|So you can go to this website, you enter some question, you get responses from two models, and you don't know what models they were generated from, and you pick the winner.| |
llm_intro_69.wav|And then depending on who wins and who loses, you can calculate the ELO scores. So the higher, the better. So what you see here is that crowding up on the top, you have the proprietary models.| |
llm_intro_70.wav|These are closed models. You don't have access to the weights. They are usually behind a web interface. And this is GPT series from OpenAI and the Cloud series from Anthropic.| |
llm_intro_71.wav|But roughly speaking, what you're seeing today in the ecosystem is that the closed models work a lot better, but you can't really work with them, fine tune them, download them, et cetera.| |
llm_intro_72.wav|And all of this stuff works worse, but depending on your application, that might be good enough. And so currently I would say the open source ecosystem is trying to boost performance| |
llm_intro_73.wav|Okay, so now I'm going to switch gears and we're going to talk about the language models, how they're improving, and where all of it is going in terms of those improvements.| |
llm_intro_74.wav|So if you train a bigger model on more text, we have a lot of confidence that the next word prediction task will improve. So algorithmic progress is not necessary.| |
llm_intro_75.wav|And algorithmic progress is kind of like a nice bonus and a lot of these organizations invest a lot into it. But fundamentally the scaling kind of offers one guaranteed path to success.| |
llm_intro_76.wav|And instead of speaking in abstract terms, I'd like to work with a concrete example that we can sort of step through. So I went to ChessGPT and I gave the following query.| |
llm_intro_77.wav|So in this case, a very reasonable tool to use would be, for example, the browser. So if you and I were faced with the same problem, you would probably go off and you would do a search, right?| |
llm_intro_78.wav|And that's exactly what ChachiPT does. So it has a way of emitting special words that we can sort of look at and we can basically look at it trying to perform a search.| |
llm_intro_79.wav|It works very similar to how you and I would do research using browsing. And it organizes this into the following information. And it sort of responds in this way.| |
llm_intro_80.wav|So it's collected the information. We have a table. We have series A, B, C, D, and E. We have the date, the amount raised, and the implied valuation in the series.| |
llm_intro_81.wav|On the bottom, it said that, actually, I apologize, I was not able to find the series A and B valuations. It only found the amounts raised. So you see how there's a not available in the table.| |
llm_intro_82.wav|So, okay, we can now continue this kind of interaction. So I said, okay, let's try to guess or impute the valuation for series A and B based on the ratios we see in series C, D, and E.| |
llm_intro_83.wav|That would be very complicated because you and I are not very good at math. In the same way, ChachiPT, just in its head sort of, is not very good at math either.| |
llm_intro_84.wav|I'm saying the x-axis is the date and the y-axis is the valuation of ScaleAI. Use logarithmic scale for y-axis, make it very nice, professional, and use gridlines.| |
llm_intro_85.wav|And so now we're looking at this and we'd like to do more tasks. So for example, let's now add a linear trend line to this plot, and we'd like to extrapolate the valuation to the end of 2025.| |
llm_intro_86.wav|And ChatGPT goes off, writes all of the code, not shown, and sort of gives the analysis. So on the bottom, we have the date, we've extrapolated, and this is the valuation.| |
llm_intro_87.wav|So based on this fit, today's valuation is $150 billion, apparently, roughly. And at the end of 2025, Scale.ai is expected to be a $2 trillion company. So congratulations to the team.| |
llm_intro_88.wav|In this case, this tool is DALI, which is also a tool developed by OpenAI. It takes natural language descriptions and it generates images. Here, DALI was used as a tool to generate this image.| |
llm_intro_89.wav|We use tons of tools, we find computers very useful, and the exact same is true for large language models, and this is increasingly a direction that is utilized by these models.| |
llm_intro_90.wav|Okay, so I've shown you here that ChatGPT can generate images. Now, multimodality is actually like a major axis along which large language models are getting better.| |
llm_intro_91.wav|So in this famous demo from Greg Brockman, one of the founders of OpenAI, he showed ChatGPT a picture of a little MyJoke website diagram that he just, you know, sketched out with a pencil.| |
llm_intro_92.wav|You can go to this MyJoke website, and you can see a little joke, and you can click to reveal a punchline. And this just works. So it's quite remarkable that this works.| |
llm_intro_93.wav|And fundamentally, you can basically start plugging images into the language models alongside with text. And ChatJPT is able to access that information and utilize it.| |
llm_intro_94.wav|Now, I mentioned that the major axis here is multimodality, so it's not just about images, seeing them and generating them, but also, for example, about audio. So, ChatGPT can now both hear and speak.| |
llm_intro_95.wav|Okay, so now I would like to switch gears to talking about some of the future directions of development in larger language models that the field broadly is interested in.| |
llm_intro_96.wav|It's just some of the things that people are thinking about. The first thing is this idea of system 1 versus system 2 type of thinking that was popularized by this book, Thinking Fast and Slow.| |
llm_intro_97.wav|So what is the distinction? The idea is that your brain can function in two kind of different modes. The system 1 thinking is your quick, instinctive, and automatic sort of part of the brain.| |
llm_intro_98.wav|So for example, if I ask you, what is 2 plus 2? You're not actually doing that math. You're just telling me it's 4, because it's available. It's cached. It's instinctive.| |
llm_intro_99.wav|You have to work out the problem in your head and give the answer. Another example is if some of you potentially play chess, when you're doing speed chess, you don't have time to think.| |
End of preview. Expand
in Dataset Viewer.
No dataset card yet
New: Create and edit this dataset card directly on the website!
Contribute a Dataset Card- Downloads last month
- 5