CLIP_benchmark

Model comparator

Evaluates and compares the performance of various CLIP-like models on different tasks and datasets.

CLIP-like model evaluation

GitHub

632 stars
12 watching
80 forks
Language: Jupyter Notebook
last commit: 5 months ago

Related projects:

Repository Description Stars
laion-ai/clap A library for learning audio embeddings from text and audio data using contrastive language-audio pretraining 1,457
laion-ai/laion-datasets A repository containing a collection of large datasets used for training and testing AI models, specifically designed to improve image-text matching capabilities. 239
catboost/benchmarks Comparative benchmarks of various machine learning algorithms 169
cloud-cv/evalai A platform for comparing and evaluating AI and machine learning algorithms at scale 1,779
pair-code/llm-comparator Analyzes LLM responses side-by-side to compare and contrast differences in generated text 347
damo-nlp-sg/m3exam A benchmark for evaluating large language models in multiple languages and formats 93
xml-comp/xml-comp Automates comparison and synchronization of XML documents across directories. 21
neulab/explainaboard An interactive tool to analyze and compare the performance of natural language processing models 362
multimodal-art-projection/omnibench Evaluates and benchmarks multimodal language models' ability to process visual, acoustic, and textual inputs simultaneously. 15
pku-yuangroup/video-bench Evaluates and benchmarks large language models' video understanding capabilities 121
01-ai/yi A series of large language models trained from scratch to excel in multiple NLP tasks 7,743
faceperceiver/laion-face Provides pre-trained face detection and analysis models using large-scale image-text data 281
aifeg/benchlmm An open-source benchmarking framework for evaluating cross-style visual capability of large multimodal models 84
lge-arc-advancedai/auptimizer Automates model building and deployment process by optimizing hyperparameters and compressing models for edge computing. 200
qcri/llmebench A benchmarking framework for large language models 81