{"height":"190","provider_name":"Hatena Blog","provider_url":"https://hatena.blog","html":"<iframe src=\"https://hatenablog-parts.com/embed?url=https%3A%2F%2Fen.bioerrorlog.work%2Fentry%2Fgpt-2-paper\" title=\"Understanding GPT-2 | Paper Summary: Language Models are Unsupervised Multitask Learners - BioErrorLog Tech Blog\" class=\"embed-card embed-blogcard\" scrolling=\"no\" frameborder=\"0\" style=\"display: block; width: 100%; height: 190px; max-width: 500px; margin: 10px 0px;\"></iframe>","description":"This is a summary of the GPT-2 paper \"Language Models are Unsupervised Multitask Learners.\" Introduction Language Models are Unsupervised Multitask Learners Overview Method Creating the WebText Training Dataset BPE: Byte Pair Encoding Model Architecture Results Language Modeling Tasks Common Sense R\u2026","blog_url":"https://en.bioerrorlog.work/","type":"rich","title":"Understanding GPT-2 | Paper Summary: Language Models are Unsupervised Multitask Learners","url":"https://en.bioerrorlog.work/entry/gpt-2-paper","width":"100%","blog_title":"BioErrorLog Tech Blog","version":"1.0","author_url":"https://blog.hatena.ne.jp/BioErrorLog/","categories":["AI","LLM","Papers"],"published":"2025-12-27 11:31:17","author_name":"BioErrorLog","image_url":"https://cdn-ak.f.st-hatena.com/images/fotolife/B/BioErrorLog/20240502/20240502164851.png"}