VBench

Video model evaluator

A benchmark suite for evaluating the performance of video generative models

[CVPR2024 Highlight] VBench - We Evaluate Video Generation

GitHub

643 stars
11 watching
33 forks
Language: Python
last commit: about 1 month ago
Linked from 1 awesome list

aigcbenchmarkdatasetevaluation-kitgen-aistable-diffusiontext-to-videovideo-generation

Backlinks from these awesome lists:

Related projects:

Repository Description Stars
cvondrick/vatic Tools for efficiently scaling up video annotation using crowdsourced marketplaces. 609
cloud-cv/evalai A platform for comparing and evaluating AI and machine learning algorithms at scale 1,779
pku-yuangroup/video-bench Evaluates and benchmarks large language models' video understanding capabilities 121
open-compass/vlmevalkit An evaluation toolkit for large vision-language models 1,514
shi-labs/vcoder An adapter for improving large language models at object-level perception tasks with auxiliary perception modalities 266
chenllliang/mmevalpro A benchmarking framework for evaluating Large Multimodal Models by providing rigorous metrics and an efficient evaluation pipeline. 22
tsb0601/mmvp An evaluation framework for multimodal language models' visual capabilities using image and question benchmarks. 296
vchahun/gv-crawl Automates text extraction and alignment from Global Voices articles to create parallel corpora for low-resource languages. 9
ytsvetko/qvec A tool to measure the quality of word vectors based on their correlation with manually crafted lexical resources 75
shangwei5/vidue A deep learning model that jointly performs video frame interpolation and deblurring with unknown exposure time 69
viame/viame A comprehensive computer vision toolkit with tools and algorithms for video and image analytics in multiple environments. 291
tianyi-lab/hallusionbench An image-context reasoning benchmark designed to challenge large vision-language models and help improve their accuracy 259
vivisect/vivisect A framework for reverse-engineering and analyzing binary executables by disassembling and emulating them. 944
huggingface/evaluate An evaluation framework for machine learning models and datasets, providing standardized metrics and tools for comparing model performance. 2,063
ofa-sys/touchstone A tool to evaluate vision-language models by comparing their performance on various tasks such as image recognition and text generation. 79