4:44 GPT: "generative pre-trained transformers"
6:39 self-attention mechanism - allows the model to focus on the most important parts of the input; adjusting importance of other words based on a keyword in the input, which is an understanding of context
No comments:
Post a Comment