Caption-Anything
Captioner
A tool generating descriptive captions from images with customizable controls and text styles.
Caption-Anything is a versatile tool combining image segmentation, visual captioning, and ChatGPT, generating tailored captions with diverse controls for user preferences. https://huggingface.co/spaces/TencentARC/Caption-Anything https://huggingface.co/spaces/VIPLab/Caption-Anything
2k stars
16 watching
103 forks
Language: Python
last commit: over 1 year ago chatgptcontrollable-generationcontrollable-image-captioningimage-captioningsegment-anything
Related projects:
Repository | Description | Stars |
---|---|---|
| An unsupervised image captioning framework that allows generating captions from images without paired data. | 215 |
| An image caption generation system using a neural network architecture with pre-trained models. | 64 |
| A library to generate images with distorted text and background patterns for security purposes. | 8 |
| A PyTorch-based tool for generating captions from images | 128 |
| Automated captioning and transcription tool for video and audio files | 74 |
| A PyTorch implementation of a framework for generating captions with styles for images and videos. | 63 |
| Trains image paragraph captioning models to generate diverse and accurate captions | 90 |
| PyTorch implementation of video captioning, combining deep learning and computer vision techniques. | 402 |
| An approach to image captioning that leverages the CLIP model and fine-tunes a language model without requiring additional supervision or object annotation. | 1,326 |
| An image caption generation model that uses abstract scene graphs to fine-grained control and generate captions | 200 |
| Enables automatic generation of descriptive text from images and videos based on user input. | 457 |
| A PyTorch implementation of image captioning models via scene graph decomposition. | 96 |
| Enhances language models to generate text based on visual descriptions of images | 352 |
| An implementation of a dense video captioning model with attention-based fusion and context gating | 149 |
| This project provides an official PyTorch implementation of a method to interpret and edit vision-language representations to mitigate hallucinations in image captions. | 46 |