 BERT-CCPoem
 BERT-CCPoem 
 Poetry model
 A BERT-based pre-trained model for Chinese classical poetry
BERT-CCPoem is an BERT-based pre-trained model particularly for Chinese classical poetry
146 stars
 3 watching
 18 forks
 
Language: Python 
last commit: over 3 years ago   bertpoetrypretrain 
 Related projects:
| Repository | Description | Stars | 
|---|---|---|
|  | A Word-based Chinese BERT model trained on large-scale text data using pre-trained models as a foundation | 460 | 
|  | A pre-trained BERT model designed to facilitate NLP research and development with limited Thai language resources | 6 | 
|  | Provides pre-trained models for Chinese language tasks with improved performance and smaller model sizes compared to existing models. | 806 | 
|  | Provides pre-trained language models derived from Wikipedia texts for natural language processing tasks | 34 | 
|  | Trains a large Chinese language model on massive data and provides a pre-trained model for downstream tasks | 230 | 
|  | Develops lightweight yet powerful pre-trained models for natural language processing tasks | 533 | 
|  | A BERT model trained on scientific text for natural language processing tasks | 1,532 | 
|  | Develops a pre-trained language model to learn semantic knowledge from permuted text without mask labels | 356 | 
|  | Provides pre-trained Chinese language models based on the ELECTRA framework for natural language processing tasks | 1,405 | 
|  | Pre-trained language model for classical Chinese texts using RoBERTa architecture | 511 | 
|  | A repository of pre-trained language models for natural language processing tasks in Chinese | 977 | 
|  | A Polish BERT-based language model trained on various corpora for natural language processing tasks | 70 | 
|  | A pre-trained language model for multiple natural language processing tasks with support for few-shot learning and transfer learning. | 656 | 
|  | Improves pre-trained Chinese language models by incorporating a correction task to alleviate inconsistency issues with downstream tasks | 646 | 
|  | Develops and trains a large-scale, parameterized model for legal question answering and text generation | 105 |