VBench
Video model evaluator
A benchmark suite for evaluating the performance of video generative models
[CVPR2024 Highlight] VBench - We Evaluate Video Generation
643 stars
11 watching
33 forks
Language: Python
last commit: about 1 month ago
Linked from 1 awesome list
aigcbenchmarkdatasetevaluation-kitgen-aistable-diffusiontext-to-videovideo-generation
Related projects:
Repository | Description | Stars |
---|---|---|
cvondrick/vatic | Tools for efficiently scaling up video annotation using crowdsourced marketplaces. | 609 |
cloud-cv/evalai | A platform for comparing and evaluating AI and machine learning algorithms at scale | 1,779 |
pku-yuangroup/video-bench | Evaluates and benchmarks large language models' video understanding capabilities | 121 |
open-compass/vlmevalkit | An evaluation toolkit for large vision-language models | 1,514 |
shi-labs/vcoder | An adapter for improving large language models at object-level perception tasks with auxiliary perception modalities | 266 |
chenllliang/mmevalpro | A benchmarking framework for evaluating Large Multimodal Models by providing rigorous metrics and an efficient evaluation pipeline. | 22 |
tsb0601/mmvp | An evaluation framework for multimodal language models' visual capabilities using image and question benchmarks. | 296 |
vchahun/gv-crawl | Automates text extraction and alignment from Global Voices articles to create parallel corpora for low-resource languages. | 9 |
ytsvetko/qvec | A tool to measure the quality of word vectors based on their correlation with manually crafted lexical resources | 75 |
shangwei5/vidue | A deep learning model that jointly performs video frame interpolation and deblurring with unknown exposure time | 69 |
viame/viame | A comprehensive computer vision toolkit with tools and algorithms for video and image analytics in multiple environments. | 291 |
tianyi-lab/hallusionbench | An image-context reasoning benchmark designed to challenge large vision-language models and help improve their accuracy | 259 |
vivisect/vivisect | A framework for reverse-engineering and analyzing binary executables by disassembling and emulating them. | 944 |
huggingface/evaluate | An evaluation framework for machine learning models and datasets, providing standardized metrics and tools for comparing model performance. | 2,063 |
ofa-sys/touchstone | A tool to evaluate vision-language models by comparing their performance on various tasks such as image recognition and text generation. | 79 |