|
297 | 297 | " - In autoencoding task, the model adjusts its parameters to minimize the discrepancy between the input and reconstructed output, typically using a reconstruction loss such as mean squared error (MSE) or binary cross-entropy.\n", |
298 | 298 | " - The model adjusts its parameters to minimize the discrepancy between the predicted tokens and the original tokens, typically using a cross-entropy loss.\n", |
299 | 299 | "\n", |
300 | | - "### **2. T5 (Text to Text Transfer Transformer)**\n", |
| 300 | + "\n", |
| 301 | + "### **2. GPT (Generative Pre-Trained Transformer)**\n", |
| 302 | + "\n", |
| 303 | + "1. By OpenAI - Autoregressive Language Model\n", |
| 304 | + "2. **[Click Here](https://cdn.openai.com/research-covers/language-unsupervised/language_understanding_paper.pdf)** to read the original paper from OpenAI - Improving Language Understanding by Generative Pre-Training.\n", |
| 305 | + "3. Pretrained on: Proprietary Data (Data for which the rights of ownership are restricted so that the ability to freely distribute the is limited)\n", |
| 306 | + "4. Autoregressive Language Model that uses attention to predict the next token in a sequence based on the previous tokens.\n", |
| 307 | + "5. GPT relies on the decoder portion of the Transformer and ignores the encoder to become exceptionally good at generating text one token at a time.\n", |
| 308 | + "\n", |
| 309 | + "\n", |
| 310 | + "### **3. T5 (Text to Text Transfer Transformer)**\n", |
301 | 311 | "<img style=\"float: right;\" width=\"400\" height=\"400\" src=\"data/images/t5.jpeg\">\n", |
302 | 312 | "\n", |
303 | 313 | "1. In 2019, By Google - Combination of Autoencoder and Autoregressor Language Model.\n", |
|
306 | 316 | "4. T5 uses both encoder and decoder of the Transformer to become highly versatile in both processing and generating text.\n", |
307 | 317 | "5. T5 based models can generate wide range of NLP tasks, from text classification to generation.\n", |
308 | 318 | "\n", |
309 | | - "### **3. GPT (Generative Pre-Trained Transformer)**\n", |
310 | | - "\n", |
311 | | - "1. By OpenAI - Autoregressive Language Model.\n", |
312 | | - "2. Pretrained on: Proprietary Data (Data for which the rights of ownership are restricted so that the ability to freely distribute the is limited)\n", |
313 | | - "3. Autoregressive Language Model that uses attention to predict the next token in a sequence based on the previous tokens.\n", |
314 | | - "4. GPT relies on the decoder portion of the Transformer and ignores the encoder to become exceptionally good at generating text one token at a time.\n", |
315 | 319 | "\n", |
316 | 320 | "### **4. Domain Specific LLMs**\n", |
317 | 321 | "\n", |
|
0 commit comments