XVERSE-MoE-A36B

Multilingual Model

Develops and publishes large multilingual language models with advanced mixing-of-experts architecture.

XVERSE-MoE-A36B: A multilingual large language model developed by XVERSE Technology Inc.

GitHub

37 stars
3 watching
2 forks
Language: Python
last commit: 5 months ago

Related projects:

Repository Description Stars
xverse-ai/xverse-13b A large language model developed to support multiple languages and applications 648
xverse-ai/xverse-moe-a4.2b Developed by XVERSE Technology Inc. as a multilingual large language model with a unique mixture-of-experts architecture and fine-tuned for various tasks such as conversation, question answering, and natural language understanding. 36
xverse-ai/xverse-65b A large language model developed by XVERSE Technology Inc. using transformer architecture and fine-tuned on diverse data sets for various applications. 132
xverse-ai/xverse-7b A multilingual large language model developed by XVERSE Technology Inc. 50
xverse-ai/xverse-v-13b A large multimodal model for visual question answering, trained on a dataset of 2.1B image-text pairs and 8.2M instruction sequences. 78
ieit-yuan/yuan2.0-m32 A high-performance language model designed to excel in tasks like natural language understanding, mathematical computation, and code generation 182
microsoft/unicoder This repository provides pre-trained models and code for understanding and generation tasks in multiple languages. 89
eleutherai/polyglot Large language models designed to perform well in multiple languages and address performance issues with current multilingual models. 476
langboat/mengzi3 An 8B and 13B language model based on the Llama architecture with multilingual capabilities. 2,031
sergioburdisso/pyss3 A Python package implementing an interpretable machine learning model for text classification with visualization tools 336
yfzhang114/slime Develops large multimodal models for high-resolution understanding and analysis of text, images, and other data types. 143
pku-yuangroup/moe-llava A large vision-language model using a mixture-of-experts architecture to improve performance on multi-modal learning tasks 2,023
nanbeige/nanbeige Develops large language models for text understanding and generation tasks. 85
skyworkai/skywork-moe A high-performance mixture-of-experts model with innovative training techniques for language processing tasks 126
tiger-ai-lab/uniir Trains and evaluates a universal multimodal retrieval model to perform various information retrieval tasks. 114