Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Language Models are Unsupervised Multitask Learners #2

Open
ysasano opened this issue Mar 4, 2019 · 0 comments
Open

Language Models are Unsupervised Multitask Learners #2

ysasano opened this issue Mar 4, 2019 · 0 comments

Comments

@ysasano
Copy link
Owner

ysasano commented Mar 4, 2019

一言でいうと

(WikipediaとBooksCorpusしか使っていない)BERTとは異なり、Common Crawlの多様で大規模な学習データで言語モデルを学習することで、多くのタスクで"Zero-shot"で SOTAのスコアを達成した手法。「特に凝ったことをしなくても言語モデル自体が本質的にマルチタスク学習器である」ことを主張している。

論文リンク

https://blog.openai.com/better-language-models/

著者/所属機関

Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever

  • OpenAI

投稿日付(yyyy/MM/dd)

2019/2/14

概要

新規性・差分

手法

結果

コメント

ソースコード
https://github.com/openai/gpt-2

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

1 participant