Skip to content

jayl-ee/KoGPT2-Summarization

 
 

Repository files navigation

KoGPT2-Summarization

모델

한국어 GPT-2 from SKT-AI

Pre-train 된 GPT-2를 문서 요약 데이터로 2차 pre-training

embedding size: 768
hidden size: 768
max input length: 1024
vocab size: 50000
number of layers: 12
number of attention heads: 12

학습 데이터

  • 국립국어원
    • 신문 기사 요약 데이터
  • 13,167 문장

학습

Nvidia 2080ti 8장으로 학습

training lossvalidation rogue

About

Summarization using KoGPT-2

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Jupyter Notebook 99.6%
  • Python 0.4%