WoBERT
Chinese Text Model
A pre-trained Chinese language model that uses word embeddings and is designed to process Chinese text
以词为基本单位的中文BERT
458 stars
8 watching
70 forks
Language: Python
last commit: about 3 years ago Related projects:
Repository | Description | Stars |
---|---|---|
ymcui/macbert | Improves pre-trained Chinese language models by incorporating a correction task to alleviate inconsistency issues with downstream tasks | 645 |
zhuiyitechnology/roformer-sim | An upgraded version of SimBERT model with integrated retrieval and generation capabilities | 438 |
brightmart/xlnet_zh | Trains a large Chinese language model on massive data and provides a pre-trained model for downstream tasks | 230 |
ethan-yt/guwenbert | A pre-trained language model for classical Chinese based on RoBERTa and ancient literature. | 506 |
ymcui/chinese-mobilebert | An implementation of MobileBERT, a pre-trained language model, in Python for NLP tasks. | 80 |
turkunlp/wikibert | Provides pre-trained language models derived from Wikipedia texts for natural language processing tasks | 34 |
yunwentechnology/unilm | This project provides pre-trained models for natural language understanding and generation tasks using the UniLM architecture. | 438 |
zhuiyitechnology/pretrained-models | A collection of pre-trained language models for natural language processing tasks | 987 |
tchayintr/thbert | A pre-trained BERT model designed to facilitate NLP research and development with limited Thai language resources | 6 |
zhuiyitechnology/t5-pegasus | Chinese generation model based on T5 architecture, trained using PEGASUS method | 555 |
thunlp-aipoet/bert-ccpoem | A BERT-based pre-trained model for Chinese classical poetry | 145 |
ymcui/pert | Develops a pre-trained language model to learn semantic knowledge from permuted text without mask labels | 354 |
ieit-yuan/yuan2.0-m32 | A high-performance language model designed to excel in tasks like natural language understanding, mathematical computation, and code generation | 180 |
cluebenchmark/cluepretrainedmodels | Provides pre-trained models for Chinese language tasks with improved performance and smaller model sizes compared to existing models. | 804 |
zhuiyitechnology/gau-alpha | An implementation of a Gated Attention Unit-based Transformer model for natural language processing tasks | 96 |