Tencent-Hunyuan-Large

Language Model

This project makes a large language model accessible for research and development

GitHub

1k stars
25 watching
52 forks
Language: Python
last commit: 4 days ago

Related projects:

Repository Description Stars
brightmart/xlnet_zh Trains a large Chinese language model on massive data and provides a pre-trained model for downstream tasks 230
yunwentechnology/unilm This project provides pre-trained models for natural language understanding and generation tasks using the UniLM architecture. 438
shawn-ieitsystems/yuan-1.0 Large-scale language model with improved performance on NLP tasks through distributed training and efficient data processing 591
ieit-yuan/yuan2.0-m32 A high-performance language model designed to excel in tasks like natural language understanding, mathematical computation, and code generation 180
clue-ai/chatyuan Large language model for dialogue support in multiple languages 1,902
tencentarc-qq/qa-clip Provides Chinese language models with high performance for image-text retrieval and classification tasks. 48
felixgithub2017/mmcu Evaluates the semantic understanding capabilities of large Chinese language models using a multimodal dataset. 87
nkcs-iclab/linglong A pre-trained Chinese language model with a modest parameter count, designed to be accessible and useful for researchers with limited computing resources. 17
ymcui/chinese-xlnet Provides pre-trained models for Chinese natural language processing tasks using the XLNet architecture 1,653
pku-yuangroup/video-bench Evaluates and benchmarks large language models' video understanding capabilities 117
thudm/chinese-transformer-xl A pre-trained Chinese language model based on the Transformer-XL architecture. 218
bilibili/index-1.9b A lightweight, multilingual language model with a long context length 904
microsoft/unicoder This repository provides pre-trained models and code for understanding and generation tasks in multiple languages. 88
duxiaoman-di/xuanyuan Develops and releases large language models for financial applications with improved performance and features 1,067
hit-scir/chinese-mixtral-8x7b An implementation of a large language model for Chinese text processing, focusing on MoE (Multi-Headed Attention) architecture and incorporating a vast vocabulary. 641