InternLM-techreport
Exam taker
An evaluation of a multilingual large language model's capabilities on comprehensive exams and comparison with other models.
901 stars
23 watching
25 forks
last commit: over 2 years ago Related projects:
| Repository | Description | Stars |
|---|---|---|
| | A tutorial project for exploring large language models and their applications in natural language processing tasks. | 1,593 |
| | Develops a large language model to aid in Chinese medicine diagnosis and prescription recommendations. | 127 |
| | Evaluates and compares the performance of multimodal large language models on various tasks | 56 |
| | A repository of papers and resources for evaluating large language models. | 1,450 |
| | A lightweight framework for building agent-based applications using LLMs and transformer architectures | 1,924 |
| | Enables users to engage with multiple large language models simultaneously and access their APIs | 256 |
| | An implementation of a plug-and-play language model that allows users to steer the topic and attributes of large language models. | 1,132 |
| | Exploring various LLMs and their applications in natural language processing and related areas | 1,854 |
| | An open-source benchmark and evaluation tool for assessing multimodal large language models' performance in embodied decision-making tasks | 99 |
| | Korea University Large Language Model developed by researchers at Korea University and HIAI Research Institute. | 576 |
| | An LLM-free benchmark suite for evaluating MLLMs' hallucination capabilities in various tasks and dimensions | 98 |
| | A benchmark for evaluating large language models' ability to process multimodal input | 322 |
| | An automated exercises assessment platform using code grading and pluggable interfaces with existing LMS | 210 |
| | A benchmarking suite for multimodal in-context learning models | 31 |
| | Develops and releases large language models trained on vast amounts of data for various applications, including natural language understanding, text generation, and more. | 864 |