XVERSE-MoE-A36B
Multilingual Model
Develops and publishes large multilingual language models with advanced mixing-of-experts architecture.
XVERSE-MoE-A36B: A multilingual large language model developed by XVERSE Technology Inc.
37 stars
3 watching
2 forks
Language: Python
last commit: 5 months ago Related projects:
Repository | Description | Stars |
---|---|---|
| A large language model developed to support multiple languages and applications | 648 |
| Developed by XVERSE Technology Inc. as a multilingual large language model with a unique mixture-of-experts architecture and fine-tuned for various tasks such as conversation, question answering, and natural language understanding. | 36 |
| A large language model developed by XVERSE Technology Inc. using transformer architecture and fine-tuned on diverse data sets for various applications. | 132 |
| A multilingual large language model developed by XVERSE Technology Inc. | 50 |
| A large multimodal model for visual question answering, trained on a dataset of 2.1B image-text pairs and 8.2M instruction sequences. | 78 |
| A high-performance language model designed to excel in tasks like natural language understanding, mathematical computation, and code generation | 182 |
| This repository provides pre-trained models and code for understanding and generation tasks in multiple languages. | 89 |
| Large language models designed to perform well in multiple languages and address performance issues with current multilingual models. | 476 |
| An 8B and 13B language model based on the Llama architecture with multilingual capabilities. | 2,031 |
| A Python package implementing an interpretable machine learning model for text classification with visualization tools | 336 |
| Develops large multimodal models for high-resolution understanding and analysis of text, images, and other data types. | 143 |
| A large vision-language model using a mixture-of-experts architecture to improve performance on multi-modal learning tasks | 2,023 |
| Develops large language models for text understanding and generation tasks. | 85 |
| A high-performance mixture-of-experts model with innovative training techniques for language processing tasks | 126 |
| Trains and evaluates a universal multimodal retrieval model to perform various information retrieval tasks. | 114 |