MacBERT
Chinese language model fine-tuning tool
Improves pre-trained Chinese language models by incorporating a correction task to alleviate inconsistency issues with downstream tasks
Revisiting Pre-trained Models for Chinese Natural Language Processing (MacBERT)
645 stars
14 watching
59 forks
last commit: over 1 year ago bertlanguage-modelmacbertnlppytorchtensorflowtransformers
Related projects:
Repository | Description | Stars |
---|---|---|
ymcui/chinese-mobilebert | An implementation of MobileBERT, a pre-trained language model, in Python for NLP tasks. | 80 |
ymcui/pert | Develops a pre-trained language model to learn semantic knowledge from permuted text without mask labels | 354 |
ymcui/chinese-xlnet | Provides pre-trained models for Chinese natural language processing tasks using the XLNet architecture | 1,653 |
ymcui/chinese-electra | Provides pre-trained Chinese language models based on the ELECTRA framework for natural language processing tasks | 1,403 |
zhuiyitechnology/wobert | A pre-trained Chinese language model that uses word embeddings and is designed to process Chinese text | 458 |
ymcui/lert | A pre-trained language model designed to leverage linguistic features and outperform comparable baselines on Chinese natural language understanding tasks. | 202 |
brightmart/xlnet_zh | Trains a large Chinese language model on massive data and provides a pre-trained model for downstream tasks | 230 |
ymcui/chinese-mixtral | Develops and releases Mixtral-based models for natural language processing tasks with a focus on Chinese text generation and understanding | 584 |
langboat/mengzi | Develops lightweight yet powerful pre-trained models for natural language processing tasks | 534 |
ethan-yt/guwenbert | A pre-trained language model for classical Chinese based on RoBERTa and ancient literature. | 506 |
cluebenchmark/cluepretrainedmodels | Provides pre-trained models for Chinese language tasks with improved performance and smaller model sizes compared to existing models. | 804 |
turkunlp/wikibert | Provides pre-trained language models derived from Wikipedia texts for natural language processing tasks | 34 |
yunwentechnology/unilm | This project provides pre-trained models for natural language understanding and generation tasks using the UniLM architecture. | 438 |
cluebenchmark/electra | Trains and evaluates a Chinese language model using adversarial training on a large corpus. | 140 |
sww9370/rocbert | A pre-trained Chinese language model designed to be robust against maliciously crafted texts | 15 |