InternLM-techreport
Exam taker
An evaluation of a multilingual large language model's capabilities on comprehensive exams and comparison with other models.
901 stars
23 watching
25 forks
last commit: over 1 year ago Related projects:
Repository | Description | Stars |
---|---|---|
internlm/tutorial | A comprehensive tutorial project offering in-depth training and practice on advanced language model technologies | 1,530 |
2020meai/tcmllm | Develops a large language model to aid in Chinese medicine diagnosis and prescription recommendations. | 118 |
freedomintelligence/mllm-bench | Evaluates and compares the performance of multimodal large language models on various tasks | 55 |
mlgroupjlu/llm-eval-survey | A repository of papers and resources for evaluating large language models. | 1,433 |
internlm/lagent | A lightweight framework for building agent-based applications using LLMs and transformer architectures | 1,865 |
internlm/openaoe | Enables users to engage with multiple large language models simultaneously and access their APIs | 253 |
uber-research/pplm | An implementation of a plug-and-play language model that allows users to steer the topic and attributes of large language models. | 1,131 |
km1994/llmsninestorydemontower | Exploring various LLMs and their applications in natural language processing and related areas | 1,798 |
pkunlp-icler/pca-eval | An open-source benchmark and evaluation tool for assessing multimodal large language models' performance in embodied decision-making tasks | 100 |
nlpai-lab/kullm | Korea University Large Language Model developed by researchers at Korea University and HIAI Research Institute. | 569 |
junyangwang0410/amber | An LLM-free benchmark suite for evaluating MLLMs' hallucination capabilities in various tasks and dimensions | 93 |
ailab-cvc/seed-bench | A benchmark for evaluating large language models' ability to process multimodal input | 315 |
ucl-ingi/inginious | An automated exercises assessment platform using code grading and pluggable interfaces with existing LMS | 208 |
ys-zong/vl-icl | A benchmarking suite for multimodal in-context learning models | 28 |
vivo-ai-lab/bluelm | Develops and releases large language models trained on vast amounts of data for various applications, including natural language understanding, text generation, and more. | 852 |