CLIP_benchmark

Model comparator

Evaluates and compares the performance of various CLIP-like models on different tasks and datasets.

CLIP-like model evaluation

GitHub

615 stars
12 watching
79 forks
Language: Jupyter Notebook
last commit: 3 months ago

Related projects:

Repository Description Stars
laion-ai/clap A library for learning audio embeddings from text and audio data using contrastive language-audio pretraining 1,427
laion-ai/laion-datasets A repository containing a collection of large datasets used for training and testing AI models, specifically designed to improve image-text matching capabilities. 235
catboost/benchmarks Comparative benchmarks of various machine learning algorithms 169
cloud-cv/evalai A platform for comparing and evaluating AI and machine learning algorithms at scale 1,771
pair-code/llm-comparator Analyzes LLM responses side-by-side to compare and contrast differences in generated text 320
damo-nlp-sg/m3exam A benchmark for evaluating large language models in multiple languages and formats 92
xml-comp/xml-comp Automates comparison and synchronization of XML documents across directories. 21
neulab/explainaboard An interactive tool to analyze and compare the performance of natural language processing models 361
multimodal-art-projection/omnibench Evaluates and benchmarks multimodal language models' ability to process visual, acoustic, and textual inputs simultaneously. 14
pku-yuangroup/video-bench Evaluates and benchmarks large language models' video understanding capabilities 117
01-ai/yi A series of large language models trained from scratch to excel in multiple NLP tasks 7,719
faceperceiver/laion-face Provides pre-trained face detection and analysis models using large-scale image-text data 278
aifeg/benchlmm An open-source benchmarking framework for evaluating cross-style visual capability of large multimodal models 82
lge-arc-advancedai/auptimizer Automates model building and deployment process by optimizing hyperparameters and compressing models for edge computing. 200
qcri/llmebench A benchmarking framework for large language models 80