musiclm-pytorch
Music Generation Model
Implementation of Google's MusicLM model for music generation using attention networks and text-conditioning.
Implementation of MusicLM, Google's new SOTA model for music generation using attention networks, in Pytorch
3k stars
97 watching
258 forks
Language: Python
last commit: over 1 year ago
Linked from 1 awesome list
artificial-intelligenceattention-mechanismsdeep-learningmusic-synthesistransformers
Related projects:
Repository | Description | Stars |
---|---|---|
| Implements Google's Text-to-Image Neural Network in PyTorch using a cascading DDPM architecture with dynamic clipping and noise level conditioning. | 8,127 |
| A PyTorch module providing tools and functions for audio signal processing | 2,561 |
| Generates music based on user input prompts using the Mubert API | 2,738 |
| An implementation of DALL-E 2's text-to-image synthesis neural network in PyTorch | 11,184 |
| A deep learning library for generating high-quality audio | 21,134 |
| An implementation of RLHF on top of the PaLM architecture to enable human feedback in reinforcement learning for large language models. | 7,729 |
| Generates high-quality speech from mel-spectrograms using a flow-based network architecture | 2,294 |
| A toolkit for speaker diarization using PyTorch and speech activity detection. | 6,508 |
| A research project using machine learning to generate music and art with the goal of creating tools for artists and musicians to extend their creative processes. | 19,234 |
| Developing and pretraining a GPT-like Large Language Model from scratch | 35,405 |
| An implementation of VALL-E in PyTorch for text-to-speech synthesis | 2,970 |
| Practices implementing popular machine learning algorithms from scratch to gain a deeper understanding of their mathematics | 23,191 |
| Generates polyphonic music sequences using deep learning models and adversarial training | 28 |
| Generates polyphonic music from scratch or based on user input using a neural network architecture | 1,863 |
| A research initiative focused on developing fundamental technology to improve the performance and efficiency of large language models. | 3,747 |