Yuan-1.0

Language Model

Large-scale language model with improved performance on NLP tasks through distributed training and efficient data processing

Yuan 1.0 Large pretrained LM

GitHub

591 stars
12 watching
117 forks
Language: Python
last commit: about 9 hours ago

Related projects:

Repository Description Stars
ieit-yuan/yuan2.0-m32 A high-performance language model designed to excel in tasks like natural language understanding, mathematical computation, and code generation 180
01-ai/yi A series of large language models trained from scratch to excel in multiple NLP tasks 7,722
brightmart/xlnet_zh Trains a large Chinese language model on massive data and provides a pre-trained model for downstream tasks 230
ieit-yuan/yuan-2.0 An open-source large language model framework for building conversational AI applications 681
tencent/tencent-hunyuan-large This project makes a large language model accessible for research and development 1,144
yunwentechnology/unilm Pre-trained Chinese language model for natural language understanding and generation tasks 438
yuangongnd/ltu An audio and speech large language model implementation with pre-trained models, datasets, and inference options 390
xverse-ai/xverse-13b A large language model developed to support multiple languages and applications 649
clue-ai/chatyuan Large language model for dialogue support in multiple languages 1,903
eleutherai/polyglot Large language models designed to perform well in multiple languages and address performance issues with current multilingual models. 476
pku-yuangroup/video-bench Evaluates and benchmarks large language models' video understanding capabilities 119
nlpai-lab/kullm Korea University Large Language Model developed by researchers at Korea University and HIAI Research Institute. 570
xverse-ai/xverse-7b A multilingual large language model developed by XVERSE Technology Inc. 50
zhuiyitechnology/pretrained-models A collection of pre-trained language models for natural language processing tasks 987
pku-yuangroup/languagebind Extending pretraining models to handle multiple modalities by aligning language and video representations 735