VBench

Video model evaluator

A tool for evaluating and benchmarking video generative models in computer vision and artificial intelligence

[CVPR2024 Highlight] VBench - We Evaluate Video Generation

GitHub

576 stars
11 watching
28 forks
Language: Python
last commit: 14 days ago
Linked from 1 awesome list

aigcbenchmarkdatasetevaluation-kitgen-aistable-diffusiontext-to-videovideo-generation

Backlinks from these awesome lists:

Related projects:

Repository Description Stars
cvondrick/vatic Tools for efficiently scaling up video annotation using crowdsourced marketplaces. 607
cloud-cv/evalai A platform for comparing and evaluating AI and machine learning algorithms at scale 1,771
pku-yuangroup/video-bench Evaluates and benchmarks large language models' video understanding capabilities 117
open-compass/vlmevalkit A toolkit for evaluating large vision-language models on various benchmarks and datasets. 1,343
shi-labs/vcoder An adapter for improving large language models at object-level perception tasks with auxiliary perception modalities 261
chenllliang/mmevalpro A benchmarking framework for evaluating Large Multimodal Models by providing rigorous metrics and an efficient evaluation pipeline. 22
tsb0601/mmvp An evaluation framework for multimodal language models' visual capabilities using image and question benchmarks. 288
vchahun/gv-crawl Automates text extraction and alignment from Global Voices articles to create parallel corpora for low-resource languages. 9
ytsvetko/qvec Tools to evaluate word vectors based on their alignment with lexical resources and features. 75
shangwei5/vidue A deep learning model that jointly performs video frame interpolation and deblurring with unknown exposure time 66
viame/viame A comprehensive computer vision toolkit with tools and algorithms for video and image analytics in multiple environments. 288
tianyi-lab/hallusionbench An image-context reasoning benchmark designed to challenge large vision-language models and help improve their accuracy 243
vivisect/vivisect A framework for reverse-engineering and analyzing binary executables by disassembling and emulating them. 939
huggingface/evaluate An evaluation framework for machine learning models and datasets, providing standardized metrics and tools for comparing model performance. 2,034
ofa-sys/touchstone A tool to evaluate vision-language models by comparing their performance on various tasks such as image recognition and text generation. 78