PoLitBert
Polish LLM
Trains a language model using a RoBERTa architecture on high-quality Polish text data
Polish RoBERTA model trained on Polish literature, Wikipedia, and Oscar. The major assumption is that quality text will give a good model.
33 stars
11 watching
3 forks
Language: Python
last commit: over 3 years ago
Linked from 1 awesome list
nlppolishrobertatext-corpus
Related projects:
Repository | Description | Stars |
---|---|---|
| A Python library providing an API for sentiment analysis of Polish text using deep learning and Word2vec models | 27 |
| Trains German transformer models to improve language understanding | 23 |
| A Polish BERT-based language model trained on various corpora for natural language processing tasks | 70 |
| A BERT-based language model pre-trained on Polish corpora for understanding Polish language. | 65 |
| An implementation of DeepMind's Relational Recurrent Neural Networks (Santoro et al. 2018) in PyTorch for word language modeling | 245 |
| A Pytorch implementation of a neural network model for machine translation | 47 |
| Pre-trained language models for biomedical natural language processing tasks | 560 |
| Pre-trained models and resources for Natural Language Processing in Polish | 329 |
| A lightweight, multilingual language model with a long context length | 920 |
| Trains a large-scale PyTorch language model on the 1-Billion Word dataset | 123 |
| An open-source implementation of a vision-language instructed large language model | 513 |
| A practical course teaching large language models and their applications through hands-on projects using OpenAI API and Hugging Face library. | 1,338 |
| Trains a large Chinese language model on massive data and provides a pre-trained model for downstream tasks | 230 |
| A collection of pre-trained natural language processing models | 170 |
| Large-scale language model with improved performance on NLP tasks through distributed training and efficient data processing | 591 |