forge

Scientific LLMs

Pre-training large language models on scientific data for downstream applications

GitHub

12 stars
1 watching
3 forks
Language: Jupyter Notebook
last commit: 9 months ago

Related projects:

Repository Description Stars
01-ai/yi A series of large language models trained from scratch to excel in multiple NLP tasks 7,699
damo-nlp-sg/m3exam A benchmark for evaluating large language models in multiple languages and formats 92
rdspring1/pytorch_gbw_lm Trains a large-scale PyTorch language model on the 1-Billion Word dataset 123
gmftbygmftby/science-llm A large-scale language model for scientific domain training on redpajama arXiv split 122
xverse-ai/xverse-7b A multilingual large language model developed by XVERSE Technology Inc. 50
luogen1996/lavin An open-source implementation of a vision-language instructed large language model 508
deepseek-ai/deepseek-moe A large language model with improved efficiency and performance compared to similar models 1,006
deepseek-ai/deepseek-llm A large language model trained on a massive dataset for various applications 1,450
flagai-open/aquila2 Provides pre-trained language models and tools for fine-tuning and evaluation 437
neso613/asr_tflite Provides pre-trained ASR models for efficient inference using TFLite 11
ai-hypercomputer/maxtext A high-performance LLM written in Python/Jax for training and inference on Google Cloud TPUs and GPUs. 1,529
academic-hammer/hammerllm A large language model pre-trained on Chinese and English data, suitable for natural language processing tasks. 43
vhellendoorn/code-lms A guide to using pre-trained large language models in source code analysis and generation 1,782
nlpai-lab/kullm Korea University Large Language Model developed by researchers at Korea University and HIAI Research Institute. 569
pratyushmaini/llm_dataset_inference Detects whether a given text sequence is part of the training data used to train a large language model. 23