gerpt2
GPT2 models
Custom German language model variants of GPT2 for natural language processing tasks.
German small and large versions of GPT2.
20 stars
1 watching
0 forks
Language: Python
last commit: over 2 years ago
Linked from 1 awesome list
common-crawlgermangpt2language-modelmachine-learningnlp
Related projects:
Repository | Description | Stars |
---|---|---|
german-nlp-group/german-transformer-training | Trains German transformer models to improve language understanding | 23 |
imcaspar/gpt2-ml | A collection of pre-trained GPT2 models and training scripts for multiple languages, including Chinese. | 1,716 |
dfki-nlp/gevalm | Evaluates German transformer language models with syntactic agreement tests | 7 |
rdspring1/pytorch_gbw_lm | Trains a large-scale PyTorch language model on the 1-Billion Word dataset | 123 |
graykode/gpt-2-pytorch | An implementation of the GPT-2 language model in PyTorch for generating text | 973 |
dbmdz/berts | Provides pre-trained language models for natural language processing tasks | 155 |
agrigpts/agrigpts | Developing large-scale language models for agricultural applications to support research and practical use in the field of agriculture | 22 |
matthias-wright/flaxmodels | Provides pre-trained deep learning models for the Jax/Flax ecosystem. | 238 |
jaxgaussianprocesses/gpjax | Provides a low-level interface to Gaussian process models in JAX for flexible extension and customisation | 461 |
vhellendoorn/code-lms | A guide to using pre-trained large language models in source code analysis and generation | 1,782 |
markuskiller/textblob-de | An extension to TextBlob for German language support | 104 |
antmicro/gerber2blend | Generates 3D models from PCB fabrication Gerber files | 9 |
brightmart/xlnet_zh | Trains a large Chinese language model on massive data and provides a pre-trained model for downstream tasks | 230 |
dfm/tinygp | A lightweight library for building Gaussian Process models in Python | 296 |
microsoft/unicoder | This repository provides pre-trained models and code for understanding and generation tasks in multiple languages. | 88 |