Yi
NLP model
A series of large language models trained from scratch to excel in multiple NLP tasks
A series of large language models trained from scratch by developers @01-ai
8k stars
107 watching
485 forks
Language: Jupyter Notebook
last commit: about 2 months ago large-language-models
Related projects:
Repository | Description | Stars |
---|---|---|
shawn-ieitsystems/yuan-1.0 | Large-scale language model with improved performance on NLP tasks through distributed training and efficient data processing | 591 |
ieit-yuan/yuan2.0-m32 | A high-performance language model designed to excel in tasks like natural language understanding, mathematical computation, and code generation | 182 |
multimodal-art-projection/map-neo | A large language model designed for research and application in natural language processing tasks. | 887 |
brightmart/xlnet_zh | Trains a large Chinese language model on massive data and provides a pre-trained model for downstream tasks | 230 |
siat-nlp/hanfei | Develops and trains a large-scale, parameterized model for legal question answering and text generation | 105 |
01-ai/yi-1.5 | An artificial intelligence model designed to improve coding, math, and reasoning capabilities while maintaining language understanding | 531 |
zhuiyitechnology/pretrained-models | A collection of pre-trained language models for natural language processing tasks | 989 |
ymcui/chinese-xlnet | Provides pre-trained models for Chinese natural language processing tasks using the XLNet architecture | 1,652 |
balavenkatesh3322/nlp-pretrained-model | A collection of pre-trained natural language processing models | 170 |
bilibili/index-1.9b | A lightweight, multilingual language model with a long context length | 920 |
jd-aig/nlp_baai | A collection of natural language processing models and tools for collaboration on a joint project between BAAI and JDAI. | 254 |
juliatext/textmodels.jl | Provides practical neural network-based models for natural language processing in Julia. | 29 |
clue-ai/promptclue | A pre-trained language model for multiple natural language processing tasks with support for few-shot learning and transfer learning. | 656 |
zhuiyitechnology/gau-alpha | An implementation of a transformer-based NLP model utilizing gated attention units | 98 |