forge
Scientific LLMs
Pre-training large language models on scientific data for downstream applications
12 stars
1 watching
3 forks
Language: Jupyter Notebook
last commit: about 1 year ago Related projects:
Repository | Description | Stars |
---|---|---|
| A series of large language models trained from scratch to excel in multiple NLP tasks | 7,743 |
| A benchmark for evaluating large language models in multiple languages and formats | 93 |
| Trains a large-scale PyTorch language model on the 1-Billion Word dataset | 123 |
| A large-scale language model for scientific domain training on redpajama arXiv split | 125 |
| A multilingual large language model developed by XVERSE Technology Inc. | 50 |
| An open-source implementation of a vision-language instructed large language model | 513 |
| A large language model with improved efficiency and performance compared to similar models | 1,024 |
| A large language model trained on a massive dataset for various applications | 1,512 |
| Provides pre-trained language models and tools for fine-tuning and evaluation | 439 |
| Provides pre-trained ASR models for efficient inference using TFLite | 11 |
| A high-performance LLM written in Python/Jax for training and inference on Google Cloud TPUs and GPUs. | 1,557 |
| A large language model pre-trained on Chinese and English data, suitable for natural language processing tasks. | 43 |
| A guide to using pre-trained large language models in source code analysis and generation | 1,789 |
| Korea University Large Language Model developed by researchers at Korea University and HIAI Research Institute. | 576 |
| Detects whether a given text sequence is part of the training data used to train a large language model. | 23 |