MMCBench
Model robustness tester
A benchmarking framework designed to evaluate the robustness of large multimodal models against common corruption scenarios
27 stars
5 watching
0 forks
Language: Python
last commit: about 1 year ago Related projects:
Repository | Description | Stars |
---|---|---|
| Evaluates and compares the performance of multimodal large language models on various tasks | 56 |
| Measures the understanding of massive multitask Chinese datasets using large language models | 87 |
| Evaluates and benchmarks the robustness of deep learning models to various corruptions and perturbations in computer vision tasks. | 1,030 |
| A toolbox for researching and evaluating robustness against attacks on machine learning models | 1,311 |
| A modular framework for testing and exploiting Modbus protocol vulnerabilities in industrial control systems | 74 |
| A benchmark for evaluating large language models' ability to process multimodal input | 322 |
| A collection of benchmarks to evaluate the multi-modal understanding capability of large vision language models. | 168 |
| A standardized benchmark for measuring the robustness of machine learning models against adversarial attacks | 682 |
| A pre-trained Chinese language model designed to be robust against maliciously crafted texts | 15 |
| A benchmarking framework for evaluating Large Multimodal Models by providing rigorous metrics and an efficient evaluation pipeline. | 22 |
| Provides a framework for computing tight certificates of adversarial robustness for randomly smoothed classifiers. | 17 |
| An evaluation framework for multimodal language models' visual capabilities using image and question benchmarks. | 296 |
| Develops a large-scale dataset and benchmark for training multimodal chart understanding models using large language models. | 87 |
| A toolset to evaluate the robustness of machine learning models | 466 |
| Automated attack synthesis tool for discovering vulnerabilities in CPU architecture and cryptographic protocols | 18 |