Skip to content

Commit c1f2d3d

Browse files
committed
Update transformers_llms_and_genai.ipynb
1 parent 20c04ca commit c1f2d3d

File tree

1 file changed

+11
-7
lines changed

1 file changed

+11
-7
lines changed

Module 9 - GenAI (LLMs and Prompt Engineering)/2. Intro to Transformers, LLMs and GenAI/transformers_llms_and_genai.ipynb

Lines changed: 11 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -297,7 +297,17 @@
297297
" - In autoencoding task, the model adjusts its parameters to minimize the discrepancy between the input and reconstructed output, typically using a reconstruction loss such as mean squared error (MSE) or binary cross-entropy.\n",
298298
" - The model adjusts its parameters to minimize the discrepancy between the predicted tokens and the original tokens, typically using a cross-entropy loss.\n",
299299
"\n",
300-
"### **2. T5 (Text to Text Transfer Transformer)**\n",
300+
"\n",
301+
"### **2. GPT (Generative Pre-Trained Transformer)**\n",
302+
"\n",
303+
"1. By OpenAI - Autoregressive Language Model\n",
304+
"2. **[Click Here](https://cdn.openai.com/research-covers/language-unsupervised/language_understanding_paper.pdf)** to read the original paper from OpenAI - Improving Language Understanding by Generative Pre-Training.\n",
305+
"3. Pretrained on: Proprietary Data (Data for which the rights of ownership are restricted so that the ability to freely distribute the is limited)\n",
306+
"4. Autoregressive Language Model that uses attention to predict the next token in a sequence based on the previous tokens.\n",
307+
"5. GPT relies on the decoder portion of the Transformer and ignores the encoder to become exceptionally good at generating text one token at a time.\n",
308+
"\n",
309+
"\n",
310+
"### **3. T5 (Text to Text Transfer Transformer)**\n",
301311
"<img style=\"float: right;\" width=\"400\" height=\"400\" src=\"data/images/t5.jpeg\">\n",
302312
"\n",
303313
"1. In 2019, By Google - Combination of Autoencoder and Autoregressor Language Model.\n",
@@ -306,12 +316,6 @@
306316
"4. T5 uses both encoder and decoder of the Transformer to become highly versatile in both processing and generating text.\n",
307317
"5. T5 based models can generate wide range of NLP tasks, from text classification to generation.\n",
308318
"\n",
309-
"### **3. GPT (Generative Pre-Trained Transformer)**\n",
310-
"\n",
311-
"1. By OpenAI - Autoregressive Language Model.\n",
312-
"2. Pretrained on: Proprietary Data (Data for which the rights of ownership are restricted so that the ability to freely distribute the is limited)\n",
313-
"3. Autoregressive Language Model that uses attention to predict the next token in a sequence based on the previous tokens.\n",
314-
"4. GPT relies on the decoder portion of the Transformer and ignores the encoder to become exceptionally good at generating text one token at a time.\n",
315319
"\n",
316320
"### **4. Domain Specific LLMs**\n",
317321
"\n",

0 commit comments

Comments
 (0)