Kolors
Text-to-Image Synthesis Framework
A Python framework for training and deploying photorealistic text-to-image synthesis models.
Kolors Team
4k stars
40 watching
293 forks
Language: Python
last commit: 4 months ago Related projects:
Repository | Description | Stars |
---|---|---|
| An implementation of DALL-E 2's text-to-image synthesis neural network in PyTorch | 11,184 |
| Image restoration toolbox with training and testing codes for various deep learning-based methods | 2,994 |
| A multilingual text2image latent diffusion model with improved aesthetics and controllability | 2,774 |
| An open source JavaScript library that allows running Keras models in the browser with GPU support using WebGL. | 4,964 |
| A PyTorch-based library for training and using state-of-the-art diffusion models to generate images, audio, and 3D structures | 26,676 |
| A Python implementation of an image-to-image translation model for generating diverse images across multiple domains. | 3,513 |
| This project generates animated videos from open-domain images by leveraging pre-trained video diffusion priors. | 2,668 |
| An implementation of a deep learning network for human pose estimation using high-resolution representations | 4,354 |
| An implementation of a neural network structure to control diffusion models by adding extra conditions. | 30,944 |
| This project is a software implementation of a diffusion model architecture, allowing users to generate synthetic images based on a learned distribution. | 6,366 |
| A software framework for training and utilizing various types of diffusion models. | 6,641 |
| A PyTorch model definition and inference/sampling code repository for a powerful diffusion transformer with fine-grained Chinese understanding | 3,678 |
| An open-source software project that develops and provides tools for restoring degraded faces in images using Generative Pre-trained Networks. | 2,456 |
| Generates images from text prompts using a variant of the DALL-E model | 14,756 |
| A framework that enables efficient training of text-to-image models by compressing the computationally expensive stage into a latent space | 531 |