MacBERT

Chinese language model fine-tuning tool

Improves pre-trained Chinese language models by incorporating a correction task to alleviate inconsistency issues with downstream tasks

Revisiting Pre-trained Models for Chinese Natural Language Processing (MacBERT)

GitHub

645 stars
14 watching
59 forks
last commit: over 1 year ago
bertlanguage-modelmacbertnlppytorchtensorflowtransformers

Related projects:

Repository Description Stars
ymcui/chinese-mobilebert An implementation of MobileBERT, a pre-trained language model, in Python for NLP tasks. 80
ymcui/pert Develops a pre-trained language model to learn semantic knowledge from permuted text without mask labels 354
ymcui/chinese-xlnet Provides pre-trained models for Chinese natural language processing tasks using the XLNet architecture 1,653
ymcui/chinese-electra Provides pre-trained Chinese language models based on the ELECTRA framework for natural language processing tasks 1,403
zhuiyitechnology/wobert A pre-trained Chinese language model that uses word embeddings and is designed to process Chinese text 458
ymcui/lert A pre-trained language model designed to leverage linguistic features and outperform comparable baselines on Chinese natural language understanding tasks. 202
brightmart/xlnet_zh Trains a large Chinese language model on massive data and provides a pre-trained model for downstream tasks 230
ymcui/chinese-mixtral Develops and releases Mixtral-based models for natural language processing tasks with a focus on Chinese text generation and understanding 584
langboat/mengzi Develops lightweight yet powerful pre-trained models for natural language processing tasks 534
ethan-yt/guwenbert A pre-trained language model for classical Chinese based on RoBERTa and ancient literature. 506
cluebenchmark/cluepretrainedmodels Provides pre-trained models for Chinese language tasks with improved performance and smaller model sizes compared to existing models. 804
turkunlp/wikibert Provides pre-trained language models derived from Wikipedia texts for natural language processing tasks 34
yunwentechnology/unilm This project provides pre-trained models for natural language understanding and generation tasks using the UniLM architecture. 438
cluebenchmark/electra Trains and evaluates a Chinese language model using adversarial training on a large corpus. 140
sww9370/rocbert A pre-trained Chinese language model designed to be robust against maliciously crafted texts 15