multimodal_rerprogramming
Model reprogramming
Cross-modal Adversarial Reprogramming enables retraining of image models on text classification tasks
Multimodal adversarial rerprogramming
11 stars
5 watching
1 forks
Language: Jupyter Notebook
last commit: about 3 years ago Related projects:
Repository | Description | Stars |
---|---|---|
paarthneekhara/rnn_adversarial_reprogramming | Repurposes pre-trained neural networks for new classification tasks through adversarial reprogramming of their inputs. | 6 |
prinsphield/adversarial_reprogramming | This project enables reprogramming of pre-trained neural networks to work on new tasks by fine-tuning them on smaller datasets. | 33 |
multimodal-art-projection/omnibench | Evaluates and benchmarks multimodal language models' ability to process visual, acoustic, and textual inputs simultaneously. | 14 |
dodohow1011/speechadvreprogram | Developing low-resource speech command recognition systems using adversarial reprogramming and transfer learning | 18 |
yunyuntsai/black-box-adversarial-reprogramming | An approach to adapt machine learning models using scarce data and limited resources by modifying their internal workings without changing the model's original architecture or training data. | 37 |
pku-yuangroup/languagebind | Extending pretraining models to handle multiple modalities by aligning language and video representations | 723 |
huckiyang/voice2series-reprogramming | An approach to reprogramming acoustic models for time series classification using differential mel-spectrograms and adversarial training | 69 |
yerevann/warp | An approach to transfer learning for NLP tasks using adversarial reprogramming and word-level task-specific embeddings. | 83 |
openbmb/viscpm | A family of large multimodal models supporting multimodal conversational capabilities and text-to-image generation in multiple languages | 1,089 |
subho406/omninet | An implementation of a unified architecture for multi-modal multi-task learning using PyTorch. | 512 |
ailab-cvc/seed | An implementation of a multimodal language model with capabilities for comprehension and generation | 576 |
xverse-ai/xverse-v-13b | A large multimodal model for visual question answering, trained on a dataset of 2.1B image-text pairs and 8.2M instruction sequences. | 77 |
l0sg/relational-rnn-pytorch | An implementation of DeepMind's Relational Recurrent Neural Networks (Santoro et al. 2018) in PyTorch for word language modeling | 244 |
lyuchenyang/macaw-llm | A multi-modal language model that integrates image, video, audio, and text data to improve language understanding and generation | 1,550 |
pleisto/yuren-baichuan-7b | A multi-modal large language model that integrates natural language and visual capabilities with fine-tuning for various tasks | 72 |