VBench
Video model evaluator
A tool for evaluating and benchmarking video generative models in computer vision and artificial intelligence
[CVPR2024 Highlight] VBench - We Evaluate Video Generation
576 stars
11 watching
28 forks
Language: Python
last commit: 14 days ago
Linked from 1 awesome list
aigcbenchmarkdatasetevaluation-kitgen-aistable-diffusiontext-to-videovideo-generation
Related projects:
Repository | Description | Stars |
---|---|---|
cvondrick/vatic | Tools for efficiently scaling up video annotation using crowdsourced marketplaces. | 607 |
cloud-cv/evalai | A platform for comparing and evaluating AI and machine learning algorithms at scale | 1,771 |
pku-yuangroup/video-bench | Evaluates and benchmarks large language models' video understanding capabilities | 117 |
open-compass/vlmevalkit | A toolkit for evaluating large vision-language models on various benchmarks and datasets. | 1,343 |
shi-labs/vcoder | An adapter for improving large language models at object-level perception tasks with auxiliary perception modalities | 261 |
chenllliang/mmevalpro | A benchmarking framework for evaluating Large Multimodal Models by providing rigorous metrics and an efficient evaluation pipeline. | 22 |
tsb0601/mmvp | An evaluation framework for multimodal language models' visual capabilities using image and question benchmarks. | 288 |
vchahun/gv-crawl | Automates text extraction and alignment from Global Voices articles to create parallel corpora for low-resource languages. | 9 |
ytsvetko/qvec | Tools to evaluate word vectors based on their alignment with lexical resources and features. | 75 |
shangwei5/vidue | A deep learning model that jointly performs video frame interpolation and deblurring with unknown exposure time | 66 |
viame/viame | A comprehensive computer vision toolkit with tools and algorithms for video and image analytics in multiple environments. | 288 |
tianyi-lab/hallusionbench | An image-context reasoning benchmark designed to challenge large vision-language models and help improve their accuracy | 243 |
vivisect/vivisect | A framework for reverse-engineering and analyzing binary executables by disassembling and emulating them. | 939 |
huggingface/evaluate | An evaluation framework for machine learning models and datasets, providing standardized metrics and tools for comparing model performance. | 2,034 |
ofa-sys/touchstone | A tool to evaluate vision-language models by comparing their performance on various tasks such as image recognition and text generation. | 78 |