This repo is devoted to output a Large-scale Language Model (LLM), fine-tuned to prioritize literary qualities in translations between Korean and Chinese, emphasizing emotional depth and stylistic nuances over literal word-for-word translation. The goal is to produce translations that reflect the original's tone, style, and emotional depth, ensuring that literary and expressive aspects are conveyed effectively. And the model will focus on delivering translations that are not only accurate and rich in literary expression but also as natural as human translations.
The LLM this repo used is ChatGLM3-6b, a medium size open source LLM model trained specifically with Chinese dataset. It contains 6 billion parameters, which means it can be fine-tuned with lora in a GPU with memory around 20 gigabytes.
The fine-tuning framework we adopted is LLaMA-Factory, it is a web-ui wrapped fine-tuning framework that is quite friendly for beginners. It supports fine-tuning of many LLaMA based LLM and also ChatGLM. It contains a lot of fine-tuning dataset and also supports customized dataset with specific format.
The setting for fine-tuning is listed as below:
The checkpoints are stored in checkpoints
directory, clone the LLaMA-Factory repo and copy all files into the saves/
directory of it, then you can find checkpoints in the web-ui. It can be further fine-tuned or deployed for chat.
- Collect more data to fine-tune
- Change for larger language models or fine-tune with full parameter (when I get more GPUs)
- Find out a more efficient way to prompt LLMs to do specific machine translation tasks