From 7a439e11112ac10adfc217b1b7d2448fcd7916ea Mon Sep 17 00:00:00 2001 From: Zeyao Du Date: Fri, 6 Dec 2019 13:46:59 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=A0=E9=99=A4=E5=AF=B9=E5=BE=AE=E4=BF=A1?= =?UTF-8?q?=E7=BE=A4=E7=9A=84=E6=8F=8F=E8=BF=B0?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- README.md | 1 - 1 file changed, 1 deletion(-) diff --git a/README.md b/README.md index f97290a..ba0d91d 100644 --- a/README.md +++ b/README.md @@ -4,7 +4,6 @@ - Chinese version of GPT2 training code, using BERT tokenizer or BPE tokenizer. It is based on the extremely awesome repository from HuggingFace team [Transformers](https://github.com/huggingface/transformers). Can write poems, news, novels, or train general language models. Support char level, word level and BPE level. Support large training corpus. - 中文的GPT2训练代码,使用BERT的Tokenizer或Sentencepiece的BPE model(感谢[kangzhonghua](https://github.com/kangzhonghua)的贡献,实现BPE模式需要略微修改train.py的代码)。可以写诗,新闻,小说,或是训练通用语言模型。支持字为单位或是分词模式或是BPE模式(需要略微修改train.py的代码)。支持大语料训练。 -- 微信交流群:请见Issue第一条。 ## NEWS 11.9