CLIP_benchmark
Model comparator
Evaluates and compares the performance of various CLIP-like models on different tasks and datasets.
CLIP-like model evaluation
615 stars
12 watching
79 forks
Language: Jupyter Notebook
last commit: 3 months ago Related projects:
Repository | Description | Stars |
---|---|---|
laion-ai/clap | A library for learning audio embeddings from text and audio data using contrastive language-audio pretraining | 1,427 |
laion-ai/laion-datasets | A repository containing a collection of large datasets used for training and testing AI models, specifically designed to improve image-text matching capabilities. | 235 |
catboost/benchmarks | Comparative benchmarks of various machine learning algorithms | 169 |
cloud-cv/evalai | A platform for comparing and evaluating AI and machine learning algorithms at scale | 1,771 |
pair-code/llm-comparator | Analyzes LLM responses side-by-side to compare and contrast differences in generated text | 320 |
damo-nlp-sg/m3exam | A benchmark for evaluating large language models in multiple languages and formats | 92 |
xml-comp/xml-comp | Automates comparison and synchronization of XML documents across directories. | 21 |
neulab/explainaboard | An interactive tool to analyze and compare the performance of natural language processing models | 361 |
multimodal-art-projection/omnibench | Evaluates and benchmarks multimodal language models' ability to process visual, acoustic, and textual inputs simultaneously. | 14 |
pku-yuangroup/video-bench | Evaluates and benchmarks large language models' video understanding capabilities | 117 |
01-ai/yi | A series of large language models trained from scratch to excel in multiple NLP tasks | 7,719 |
faceperceiver/laion-face | Provides pre-trained face detection and analysis models using large-scale image-text data | 278 |
aifeg/benchlmm | An open-source benchmarking framework for evaluating cross-style visual capability of large multimodal models | 82 |
lge-arc-advancedai/auptimizer | Automates model building and deployment process by optimizing hyperparameters and compressing models for edge computing. | 200 |
qcri/llmebench | A benchmarking framework for large language models | 80 |