WoBERT

Chinese Text Model

A pre-trained Chinese language model that uses word embeddings and is designed to process Chinese text

以词为基本单位的中文BERT

GitHub

458 stars
8 watching
70 forks
Language: Python
last commit: about 3 years ago

Related projects:

Repository Description Stars
ymcui/macbert Improves pre-trained Chinese language models by incorporating a correction task to alleviate inconsistency issues with downstream tasks 645
zhuiyitechnology/roformer-sim An upgraded version of SimBERT model with integrated retrieval and generation capabilities 438
brightmart/xlnet_zh Trains a large Chinese language model on massive data and provides a pre-trained model for downstream tasks 230
ethan-yt/guwenbert A pre-trained language model for classical Chinese based on RoBERTa and ancient literature. 506
ymcui/chinese-mobilebert An implementation of MobileBERT, a pre-trained language model, in Python for NLP tasks. 80
turkunlp/wikibert Provides pre-trained language models derived from Wikipedia texts for natural language processing tasks 34
yunwentechnology/unilm This project provides pre-trained models for natural language understanding and generation tasks using the UniLM architecture. 438
zhuiyitechnology/pretrained-models A collection of pre-trained language models for natural language processing tasks 987
tchayintr/thbert A pre-trained BERT model designed to facilitate NLP research and development with limited Thai language resources 6
zhuiyitechnology/t5-pegasus Chinese generation model based on T5 architecture, trained using PEGASUS method 555
thunlp-aipoet/bert-ccpoem A BERT-based pre-trained model for Chinese classical poetry 145
ymcui/pert Develops a pre-trained language model to learn semantic knowledge from permuted text without mask labels 354
ieit-yuan/yuan2.0-m32 A high-performance language model designed to excel in tasks like natural language understanding, mathematical computation, and code generation 180
cluebenchmark/cluepretrainedmodels Provides pre-trained models for Chinese language tasks with improved performance and smaller model sizes compared to existing models. 804
zhuiyitechnology/gau-alpha An implementation of a Gated Attention Unit-based Transformer model for natural language processing tasks 96