Morizeyao/GPT2-Chinese

Chinese version of GPT2 training code, using BERT tokenizer.

PythonShellnlptext-generationtransformerchinesegpt-2
This is stars and forks stats for /Morizeyao/GPT2-Chinese repository. As of 25 Apr, 2024 this repository has 7164 stars and 1735 forks.

GPT2-Chinese Description Chinese version of GPT2 training code, using BERT tokenizer or BPE tokenizer. It is based on the extremely awesome repository from HuggingFace team Transformers. Can write poems, news, novels, or train general language models. Support char level, word level and BPE level. Support large training corpus. 中文的GPT2训练代码,使用BERT的Tokenizer或Sentencepiece的BPE model(感谢kangzhonghua的贡献,实现BPE模式需要略微修改train.py的代码)。可以写诗,新闻,小说,或是训练通用语言模型。支持字为单位或是分词模式或是BPE模式(需要略微修改train.py的代码)。支持大语料训练。 UPDATE...
Read on GithubGithub Stats Page
repotechsstarsweeklyforksweekly
simonw/advent-of-code-2022-in-rustRustShellJust116040
saitamasahil/Pixel-Launcher-ExtendedShellSmali674+736+1
4ndr34z/shellsShell3030620
betaveros/advent-of-code-2022Shell201050
asquarezone/KubernetesZoneSmartyShell702300
broadinstitute/cromwhelmSmartyPythonWDL2020
evelyneee/ellekitSwiftCC++2980350
evgenyneu/keychain-swiftSwiftShellRuby2.5k03270
vikash-kumar01/Jenkins-Docker-ProjectDockerfileShell150160+2
bjw-s/container-imagesPythonDockerfileShell14020