Awesome-LLM
LLM resources
A curated collection of papers, frameworks, and resources for training and deploying large language models.
Awesome-LLM: a curated list of Large Language Model
19k stars
378 watching
2k forks
last commit: about 14 hours ago
Linked from 4 awesome lists
Awesome-LLM / Trending LLM Projects | |||
Deep-Live-Cam | 40,757 | 4 days ago | real time face swap and one-click video deepfake with only a single image (uncensored) |
MiniCPM-V 2.6 | 12,619 | about 1 month ago | A GPT-4V Level MLLM for Single Image, Multi Image and Video on Your Phone |
GPT-SoVITS | 35,728 | 14 days ago | 1 min voice data can also be used to train a good TTS model! (few shot voice cloning) |
Awesome-LLM / Milestone Papers | |||
Attention Is All You Need | |||
Improving Language Understanding by Generative Pre-Training | |||
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding | |||
Language Models are Unsupervised Multitask Learners | |||
Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism | |||
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer | |||
ZeRO: Memory Optimizations Toward Training Trillion Parameter Models | |||
Scaling Laws for Neural Language Models | |||
Language models are few-shot learners | |||
Switch Transformers: Scaling to Trillion Parameter Models with Simple and Efficient Sparsity | |||
Evaluating Large Language Models Trained on Code | |||
On the Opportunities and Risks of Foundation Models | |||
Finetuned Language Models are Zero-Shot Learners | |||
Multitask Prompted Training Enables Zero-Shot Task Generalization | |||
GLaM: Efficient Scaling of Language Models with Mixture-of-Experts | |||
WebGPT: Browser-assisted question-answering with human feedback | |||
Improving language models by retrieving from trillions of tokens | |||
Scaling Language Models: Methods, Analysis & Insights from Training Gopher | |||
Chain-of-Thought Prompting Elicits Reasoning in Large Language Models | |||
LaMDA: Language Models for Dialog Applications | |||
Solving Quantitative Reasoning Problems with Language Models | |||
Using Deep and Megatron to Train Megatron-Turing NLG 530B, A Large-Scale Generative Language Model | |||
Training language models to follow instructions with human feedback | |||
PaLM: Scaling Language Modeling with Pathways | |||
An empirical analysis of compute-optimal large language model training | |||
OPT: Open Pre-trained Transformer Language Models | |||
Unifying Language Learning Paradigms | |||
Emergent Abilities of Large Language Models | |||
Beyond the Imitation Game: Quantifying and extrapolating the capabilities of language models | 2,868 | 4 months ago | |
Language Models are General-Purpose Interfaces | |||
Improving alignment of dialogue agents via targeted human judgements | |||
Scaling Instruction-Finetuned Language Models | |||
GLM-130B: An Open Bilingual Pre-trained Model | |||
Holistic Evaluation of Language Models | |||
BLOOM: A 176B-Parameter Open-Access Multilingual Language Model | |||
Galactica: A Large Language Model for Science | |||
OPT-IML: Scaling Language Model Instruction Meta Learning through the Lens of Generalization | |||
The Flan Collection: Designing Data and Methods for Effective Instruction Tuning | |||
LLaMA: Open and Efficient Foundation Language Models | |||
Language Is Not All You Need: Aligning Perception with Language Models | |||
Resurrecting Recurrent Neural Networks for Long Sequences | |||
PaLM-E: An Embodied Multimodal Language Model | |||
GPT-4 Technical Report | |||
Visual Instruction Tuning | |||
Pythia: A Suite for Analyzing Large Language Models Across Training and Scaling | |||
Principle-Driven Self-Alignment of Language Models from Scratch with Minimal Human Supervision | |||
PaLM 2 Technical Report | |||
RWKV: Reinventing RNNs for the Transformer Era | |||
Direct Preference Optimization: Your Language Model is Secretly a Reward Model | |||
Tree of Thoughts: Deliberate Problem Solving with Large Language Models | |||
Llama 2: Open Foundation and Fine-Tuned Chat Models | |||
Mistral 7B | |||
Mamba: Linear-Time Sequence Modeling with Selective State Spaces | |||
DeepSeek-V2: A Strong, Economical, and Efficient Mixture-of-Experts Language Model | |||
Jamba: A Hybrid Transformer-Mamba Language Model | |||
Transformers are SSMs: Generalized Models and Efficient Algorithms Through Structured State Space Duality | |||
The Llama 3 Herd of Models | |||
Awesome-LLM / LLM Leaderboard | |||
Chatbot Arena Leaderboard | a benchmark platform for large language models (LLMs) that features anonymous, randomized battles in a crowdsourced manner | ||
Open LLM Leaderboard | aims to track, rank, and evaluate LLMs and chatbots as they are released | ||
ACLUE | 23 | 8 months ago | an evaluation benchmark focused on ancient Chinese language comprehension |
AlpacaEval | An Automatic Evaluator for Instruction-following Language Models using Nous benchmark suite | ||
BeHonest | A pioneering benchmark specifically designed to assess honesty in LLMs comprehensively | ||
Berkeley Function-Calling Leaderboard | evaluates LLM's ability to call external functions/tools | ||
Chinese Large Model Leaderboard | 2,819 | 5 days ago | an expert-driven benchmark for Chineses LLMs |
CompassRank | CompassRank is dedicated to exploring the most advanced language and visual models, offering a comprehensive, objective, and neutral evaluation reference for the industry and research | ||
CompMix | a benchmark evaluating QA methods that operate over a mixture of heterogeneous input sources (KB, text, tables, infoboxes) | ||
DreamBench++ | a benchmark for evaluating the performance of large language models (LLMs) in various tasks related to both textual and visual imagination | ||
FELM | a meta-benchmark that evaluates how well factuality evaluators assess the outputs of large language models (LLMs) | ||
InfiBench | a benchmark designed to evaluate large language models (LLMs) specifically in their ability to answer real-world coding-related questions | ||
LawBench | a benchmark designed to evaluate large language models in the legal domain | ||
LLMEval | focuses on understanding how these models perform in various scenarios and analyzing results from an interpretability perspective | ||
M3CoT | a benchmark that evaluates large language models on a variety of multimodal reasoning tasks, including language, natural and social sciences, physical and social commonsense, temporal reasoning, algebra, and geometry | ||
MathEval | a comprehensive benchmarking platform designed to evaluate large models' mathematical abilities across 20 fields and nearly 30,000 math problems | ||
MixEval | a ground-truth-based dynamic benchmark derived from off-the-shelf benchmark mixtures, which evaluates LLMs with a highly capable model ranking (i.e., 0.96 correlation with Chatbot Arena) while running locally and quickly (6% the time and cost of running MMLU) | ||
MMedBench | a benchmark that evaluates large language models' ability to answer medical questions across multiple languages | ||
MMToM-QA | a multimodal question-answering benchmark designed to evaluate AI models' cognitive ability to understand human beliefs and goals | ||
OlympicArena | a benchmark for evaluating AI models across multiple academic disciplines like math, physics, chemistry, biology, and more | ||
PubMedQA | a biomedical question-answering benchmark designed for answering research-related questions using PubMed abstracts | ||
SciBench | benchmark designed to evaluate large language models (LLMs) on solving complex, college-level scientific problems from domains like chemistry, physics, and mathematics | ||
SuperBench | a benchmark platform designed for evaluating large language models (LLMs) on a range of tasks, particularly focusing on their performance in different aspects such as natural language understanding, reasoning, and generalization | ||
SuperLim | a Swedish language understanding benchmark that evaluates natural language processing (NLP) models on various tasks such as argumentation analysis, semantic similarity, and textual entailment | ||
TAT-DQA | a large-scale Document Visual Question Answering (VQA) dataset designed for complex document understanding, particularly in financial reports | ||
TAT-QA | a large-scale question-answering benchmark focused on real-world financial data, integrating both tabular and textual information | ||
VisualWebArena | a benchmark designed to assess the performance of multimodal web agents on realistic visually grounded tasks | ||
We-Math | a benchmark that evaluates large multimodal models (LMMs) on their ability to perform human-like mathematical reasoning | ||
WHOOPS! | a benchmark dataset testing AI's ability to reason about visual commonsense through images that defy normal expectations | ||
Awesome-LLM / Open LLM / Meta | |||
Llama 3.2-1|3|11|90B | |||
Llama 3.1-8|70|405B | |||
Llama 3-8|70B | |||
Llama 2-7|13|70B | |||
Llama 1-7|13|33|65B | |||
OPT-1.3|6.7|13|30|66B | |||
Awesome-LLM / Open LLM / Mistral AI | |||
Codestral-7|22B | |||
Mistral-7B | |||
Mixtral-8x7B | |||
Mixtral-8x22B | |||
Awesome-LLM / Open LLM / Google | |||
Gemma2-9|27B | |||
Gemma-2|7B | |||
RecurrentGemma-2B | 607 | 5 months ago | |
T5 | |||
Awesome-LLM / Open LLM / Apple | |||
OpenELM-1.1|3B | |||
Awesome-LLM / Open LLM / Microsoft | |||
Phi1-1.3B | |||
Phi2-2.7B | |||
Phi3-3.8|7|14B | |||
Awesome-LLM / Open LLM / AllenAI | |||
OLMo-7B | |||
Awesome-LLM / Open LLM / xAI | |||
Grok-1-314B-MoE | |||
Awesome-LLM / Open LLM / Cohere | |||
Command R-35B | |||
Awesome-LLM / Open LLM / DeepSeek | |||
DeepSeek-Math-7B | |||
DeepSeek-Coder-1.3|6.7|7|33B | |||
DeepSeek-VL-1.3|7B | |||
DeepSeek-MoE-16B | |||
DeepSeek-v2-236B-MoE | |||
DeepSeek-Coder-v2-16|236B-MOE | 2,189 | about 2 months ago | |
Awesome-LLM / Open LLM / Alibaba | |||
Qwen-1.8B|7B|14B|72B | |||
Qwen1.5-0.5B|1.8B|4B|7B|14B|32B|72B|110B|MoE-A2.7B | |||
Qwen2-0.5B|1.5B|7B|57B-A14B-MoE|72B | |||
Qwen2.5-0.5B|1.5B|3B|7B|14B|32B|72B | |||
CodeQwen1.5-7B | |||
Qwen2.5-Coder-1.5B|7B|32B | |||
Qwen2-Math-1.5B|7B|72B | |||
Qwen2.5-Math-1.5B|7B|72B | |||
Qwen-VL-7B | |||
Qwen2-VL-2B|7B|72B | |||
Qwen2-Audio-7B | |||
Awesome-LLM / Open LLM / 01-ai | |||
Yi-34B | |||
Yi1.5-6|9|34B | |||
Yi-VL-6B|34B | |||
Awesome-LLM / Open LLM / Baichuan | |||
Baichuan-7|13B | |||
Baichuan2-7|13B | |||
Awesome-LLM / Open LLM / Nvidia | |||
Nemotron-4-340B | |||
Awesome-LLM / Open LLM / BLOOM | |||
BLOOMZ&mT0 | |||
Awesome-LLM / Open LLM / Zhipu AI | |||
GLM-2|6|10|13|70B | |||
CogVLM2-19B | |||
Awesome-LLM / Open LLM / OpenBMB | |||
MiniCPM-2B | |||
OmniLLM-12B | |||
VisCPM-10B | |||
CPM-Bee-1|2|5|10B | |||
Awesome-LLM / Open LLM / RWKV Foundation | |||
RWKV-v4|5|6 | |||
Awesome-LLM / Open LLM / ElutherAI | |||
Pythia-1|1.4|2.8|6.9|12B | 2,280 | 20 days ago | |
Awesome-LLM / Open LLM / Stability AI | |||
StableLM-3B | |||
StableLM-v2-1.6|12B | |||
StableCode-3B | |||
Awesome-LLM / Open LLM / BigCode | |||
StarCoder-1|3|7B | |||
StarCoder2-3|7|15B | |||
Awesome-LLM / Open LLM / DataBricks | |||
MPT-7B | |||
DBRX-132B-MoE | |||
Awesome-LLM / Open LLM / Shanghai AI Laboratory | |||
InternLM2-1.8|7|20B | |||
InternLM-Math-7B|20B | |||
InternLM-XComposer2-1.8|7B | |||
InternVL-2|6|14|26 | |||
Awesome-LLM / LLM Data | |||
LLMDataHub | 2,635 | 12 months ago | |
IBM data-prep-kit | 290 | 4 days ago | Open-Source Toolkit for Efficient Unstructured Data Processing with Pre-built Modules and Local to Cluster Scalability |
Awesome-LLM / LLM Evaluation: | |||
lm-evaluation-harness | 6,970 | 6 days ago | A framework for few-shot evaluation of language models |
MixEval | 224 | 12 days ago | A reliable click-and-go evaluation suite compatible with both open-source and proprietary models, supporting MixEval and other benchmarks |
lighteval | 804 | 6 days ago | a lightweight LLM evaluation suite that Hugging Face has been using internally |
OLMO-eval | 310 | 22 days ago | a repository for evaluating open language models |
instruct-eval | 528 | 9 months ago | This repository contains code to quantitatively evaluate instruction-tuned models such as Alpaca and Flan-T5 on held-out tasks |
simple-evals | 1,939 | 22 days ago | Eval tools by OpenAI |
Giskard | 4,071 | 6 days ago | Testing & evaluation library for LLM applications, in particular RAGs |
LangSmith | a unified platform from LangChain framework for: evaluation, collaboration HITL (Human In The Loop), logging and monitoring LLM applications | ||
Ragas | 7,233 | 7 days ago | a framework that helps you evaluate your Retrieval Augmented Generation (RAG) pipelines |
Awesome-LLM / LLM Training Frameworks | |||
DeepSpeed | 35,463 | 7 days ago | DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective |
Megatron-DeepSpeed | 1,895 | about 1 month ago | DeepSpeed version of NVIDIA's Megatron-LM that adds additional support for several features such as MoE model training, Curriculum Learning, 3D Parallelism, and others |
torchtune | 4,320 | 4 days ago | A Native-PyTorch Library for LLM Fine-tuning |
torchtitan | 2,615 | 5 days ago | A native PyTorch Library for large model training |
NeMo Framework | 12,118 | 6 days ago | Generative AI framework built for researchers and PyTorch developers working on Large Language Models (LLMs), Multimodal Models (MMs), Automatic Speech Recognition (ASR), Text to Speech (TTS), and Computer Vision (CV) domains |
Megatron-LM | 10,562 | 6 days ago | Ongoing research training transformer models at scale |
Colossal-AI | 38,797 | 6 days ago | Making large AI models cheaper, faster, and more accessible |
BMTrain | 563 | 4 months ago | Efficient Training for Big Models |
Mesh Tensorflow | 1,592 | about 1 year ago | Mesh TensorFlow: Model Parallelism Made Easier |
maxtext | 1,529 | 4 days ago | A simple, performant and scalable Jax LLM! |
GPT-NeoX | 6,941 | 6 days ago | An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library |
Awesome-LLM / LLM Deployment | |||
SGLang | 6,082 | 4 days ago | SGLang is a fast serving framework for large language models and vision language models |
vLLM | 30,303 | 4 days ago | A high-throughput and memory-efficient inference and serving engine for LLMs |
TGI | a toolkit for deploying and serving Large Language Models (LLMs) | ||
exllama | 2,760 | about 1 year ago | A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights |
llama.cpp | 67,866 | 6 days ago | LLM inference in C/C++ |
ollama | 97,979 | 6 days ago | Get up and running with Llama 3, Mistral, Gemma, and other large language models |
Langfuse | 6,537 | 4 days ago | Open Source LLM Engineering Platform 🪢 Tracing, Evaluations, Prompt Management, Evaluations and Playground |
FastChat | 36,975 | 7 days ago | A distributed multi-model LLM serving system with web UI and OpenAI-compatible RESTful APIs |
mistral.rs | 4,466 | 5 days ago | Blazingly fast LLM inference |
MindSQL | 228 | about 1 month ago | A python package for Txt-to-SQL with self hosting functionalities and RESTful APIs compatible with proprietary as well as open source LLM |
SkyPilot | 6,801 | 5 days ago | Run LLMs and batch jobs on any cloud. Get maximum cost savings, highest GPU availability, and managed execution -- all with a simple interface |
Haystack | an open-source NLP framework that allows you to use LLMs and transformer-based models from Hugging Face, OpenAI and Cohere to interact with your own data | ||
Sidekick | 2,296 | about 1 month ago | Data integration platform for LLMs |
QA-Pilot | 183 | 3 months ago | An interactive chat project that leverages Ollama/OpenAI/MistralAI LLMs for rapid understanding and navigation of GitHub code repository or compressed file resources |
Shell-Pilot | 69 | 5 months ago | Interact with LLM using Ollama models(or openAI, mistralAI)via pure shell scripts on your Linux(or MacOS) system, enhancing intelligent system management without any dependencies |
LangChain | 94,887 | 6 days ago | Building applications with LLMs through composability |
Floom | 36 | 4 days ago | AI gateway and marketplace for developers, enables streamlined integration of AI features into products |
Swiss Army Llama | 941 | about 2 months ago | Comprehensive set of tools for working with local LLMs for various tasks |
LiteChain | 416 | 11 months ago | Lightweight alternative to LangChain for composing LLMs |
magentic | 2,048 | 4 days ago | Seamlessly integrate LLMs as Python functions |
wechat-chatgpt | 13,277 | 6 months ago | Use ChatGPT On Wechat via wechaty |
promptfoo | 4,754 | 4 days ago | Test your prompts. Evaluate and compare LLM outputs, catch regressions, and improve prompt quality |
Agenta | 1,275 | 6 days ago | Easily build, version, evaluate and deploy your LLM-powered apps |
Serge | 5,681 | 3 days ago | a chat interface crafted with llama.cpp for running Alpaca models. No API keys, entirely self-hosted! |
Langroid | 2,654 | 5 days ago | Harness LLMs with Multi-Agent Programming |
Embedchain | 22,829 | 6 days ago | Framework to create ChatGPT like bots over your dataset |
Opik | 2,121 | 6 days ago | Confidently evaluate, test, and ship LLM applications with a suite of observability tools to calibrate language model outputs across your dev and production lifecycle |
IntelliServer | 27 | 9 months ago | simplifies the evaluation of LLMs by providing a unified microservice to access and test multiple AI models |
OpenLLM | 10,051 | 10 days ago | Fine-tune, serve, deploy, and monitor any open-source LLMs in production. Used in production at for LLMs-based applications |
DeepSpeed-Mii | 1,898 | 13 days ago | MII makes low-latency and high-throughput inference, similar to vLLM powered by DeepSpeed |
Text-Embeddings-Inference | 2,838 | 16 days ago | Inference for text-embeddings in Rust, HFOIL Licence |
Infinity | 1,464 | 5 days ago | Inference for text-embeddings in Python |
TensorRT-LLM | 8,668 | 8 days ago | Nvidia Framework for LLM Inference |
FasterTransformer | 5,886 | 8 months ago | NVIDIA Framework for LLM Inference(Transitioned to TensorRT-LLM) |
Flash-Attention | 14,248 | 4 days ago | A method designed to enhance the efficiency of Transformer models |
Langchain-Chatchat | 32,060 | 4 days ago | Formerly langchain-ChatGLM, local knowledge based LLM (like ChatGLM) QA app with langchain |
Search with Lepton | 7,839 | 8 days ago | Build your own conversational search engine using less than 500 lines of code by |
Robocorp | 472 | about 2 months ago | Create, deploy and operate Actions using Python anywhere to enhance your AI agents and assistants. Batteries included with an extensive set of libraries, helpers and logging |
LMDeploy | 4,653 | 5 days ago | A high-throughput and low-latency inference and serving framework for LLMs and VLs |
Tune Studio | Playground for devs to finetune & deploy LLMs | ||
LLocalSearch | 5,671 | 3 months ago | Locally running websearch using LLM chains |
AI Gateway | 6,290 | 8 days ago | — Gateway streamlines requests to 100+ open & closed source models with a unified API. It is also production-ready with support for caching, fallbacks, retries, timeouts, loadbalancing, and can be edge-deployed for minimum latency |
talkd.ai dialog | 377 | 22 days ago | Simple API for deploying any RAG or LLM that you want adding plugins |
Wllama | 437 | 21 days ago | WebAssembly binding for llama.cpp - Enabling in-browser LLM inference |
GPUStack | 600 | 5 days ago | An open-source GPU cluster manager for running LLMs |
MNN-LLM | 8,739 | 7 days ago | -- A Device-Inference framework, including LLM Inference on device(Mobile Phone/PC/IOT) |
CAMEL | First LLM Multi-agent framework | ||
Awesome-LLM / LLM Applications | |||
AdalFlow | 2,036 | 4 days ago | AdalFlow: The library to build&auto-optimize LLM applications |
dspy | 18,802 | 4 days ago | DSPy: The framework for programming—not prompting—foundation models |
YiVal | 2,656 | 7 months ago | — Evaluate and Evolve: YiVal is an open-source GenAI-Ops tool for tuning and evaluating prompts, configurations, and model parameters using customizable datasets, evaluation methods, and improvement strategies |
Guidance | 19,096 | 10 days ago | — A handy looking Python library from Microsoft that uses Handlebars templating to interleave generation, prompting, and logical control |
LangChain | 94,887 | 6 days ago | — A popular Python/JavaScript library for chaining sequences of language model prompts |
Evidently | 5,391 | 7 days ago | — An open-source framework to evaluate, test and monitor ML and LLM-powered systems |
FLAML (A Fast Library for Automated Machine Learning & Tuning) | : A Python library for automating selection of models, hyperparameters, and other tunable choices | ||
Chainlit | — A Python library for making chatbot interfaces | ||
Guardrails.ai | — A Python library for validating outputs and retrying failures. Still in alpha, so expect sharp edges and bugs | ||
Semantic Kernel | 21,946 | 6 days ago | — A Python/C#/Java library from Microsoft that supports prompt templating, function chaining, vectorized memory, and intelligent planning |
Prompttools | 2,708 | 3 months ago | — Open-source Python tools for testing and evaluating models, vector DBs, and prompts |
Outlines | 9,481 | 11 days ago | — A Python library that provides a domain-specific language to simplify prompting and constrain generation |
Promptify | 3,266 | 8 months ago | — A small Python library for using language models to perform NLP tasks |
Scale Spellbook | — A paid product for building, comparing, and shipping language model apps | ||
PromptPerfect | — A paid product for testing and improving prompts | ||
Weights & Biases | — A paid product for tracking model training and prompt engineering experiments | ||
OpenAI Evals | 15,015 | about 2 months ago | — An open-source library for evaluating task performance of language models and prompts |
LlamaIndex | 36,776 | 6 days ago | — A Python library for augmenting LLM apps with data |
Arthur Shield | — A paid product for detecting toxicity, hallucination, prompt injection, etc | ||
LMQL | — A programming language for LLM interaction with support for typed prompting, control flow, constraints, and tools | ||
ModelFusion | 1,159 | 4 months ago | A TypeScript library for building apps with LLMs and other ML models (speech-to-text, text-to-speech, image generation) |
Flappy | 311 | 7 months ago | — Production-Ready LLM Agent SDK for Every Developer |
GPTRouter | GPTRouter is an open source LLM API Gateway that offers a universal API for 30+ LLMs, vision, and image models, with smart fallbacks based on uptime and latency, automatic retries, and streaming. Stay operational even when OpenAI is down | ||
QAnything | 11,875 | 26 days ago | A local knowledge base question-answering system designed to support a wide range of file formats and databases |
OneKE | — A bilingual Chinese-English knowledge extraction model with knowledge graphs and natural language processing technologies | ||
llm-ui | 403 | 5 months ago | A React library for building LLM UIs |
Wordware | A web-hosted IDE where non-technical domain experts work with AI Engineers to build task-specific AI agents. We approach prompting as a new programming language rather than low/no-code blocks | ||
Wallaroo.AI | Deploy, manage, optimize any model at scale across any environment from cloud to edge. Let's you go from python notebook to inferencing in minutes | ||
Dify | 51,873 | 4 days ago | An open-source LLM app development platform with an intuitive interface that streamlines AI workflows, model management, and production deployment |
LazyLLM | 1,020 | 7 days ago | An open-source LLM app for building multi-agent LLMs applications in an easy and lazy way, supports model deployment and fine-tuning |
MemFree | 1,039 | 5 days ago | Open Source Hybrid AI Search Engine, Instantly Get Accurate Answers from the Internet, Bookmarks, Notes, and Docs. Support One-Click Deployment |
unslothai | 18,151 | 7 days ago | A framework that specializes in efficient fine-tuning. On its GitHub page, you can find ready-to-use fine-tuning templates for various LLMs, allowing you to easily train your own data for free on the Google Colab cloud |
Awesome-LLM / LLM Tutorials and Courses | |||
llm-course | 39,120 | 4 months ago | Course to get into Large Language Models (LLMs) with roadmaps and Colab notebooks |
UWaterloo CS 886 | Recent Advances on Foundation Models | ||
CS25-Transformers United | |||
ChatGPT Prompt Engineering | |||
Princeton: Understanding Large Language Models | |||
CS324 - Large Language Models | |||
State of GPT | |||
A Visual Guide to Mamba and State Space Models | |||
Let's build GPT: from scratch, in code, spelled out. | |||
minbpe | Minimal, clean code for the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization | ||
femtoGPT | 834 | 2 months ago | Pure Rust implementation of a minimal Generative Pretrained Transformer |
Neurips2022-Foundational Robustness of Foundation Models | |||
ICML2022-Welcome to the "Big Model" Era: Techniques and Systems to Train and Serve Bigger Models | |||
GPT in 60 Lines of NumPy | |||
Awesome-LLM / LLM Books | |||
Generative AI with LangChain: Build large language model (LLM) apps with Python, ChatGPT, and other LLMs | it comes with a that showcases a lot of the functionality | ||
Build a Large Language Model (From Scratch) | A guide to building your own working LLM | ||
BUILD GPT: HOW AI WORKS | explains how to code a Generative Pre-trained Transformer, or GPT, from scratch | ||
Hands-On Large Language Models: Language Understanding and Generation | Explore the world of Large Language Models with over 275 custom made figures in this illustrated guide! | ||
Awesome-LLM / Great thoughts about LLM | |||
Why did all of the public reproduction of GPT-3 fail? | |||
A Stage Review of Instruction Tuning | |||
LLM Powered Autonomous Agents | |||
Why you should work on AI AGENTS! | |||
Google "We Have No Moat, And Neither Does OpenAI" | |||
AI competition statement | |||
Prompt Engineering | |||
Noam Chomsky: The False Promise of ChatGPT | |||
Is ChatGPT 175 Billion Parameters? Technical Analysis | |||
The Next Generation Of Large Language Models | |||
Large Language Model Training in 2023 | |||
How does GPT Obtain its Ability? Tracing Emergent Abilities of Language Models to their Sources | |||
Open Pretrained Transformers | |||
Scaling, emergence, and reasoning in large language models | |||
Awesome-LLM / Miscellaneous | |||
Arize-Phoenix | Open-source tool for ML observability that runs in your notebook environment. Monitor and fine tune LLM, CV and Tabular Models | ||
Emergent Mind | The latest AI news, curated & explained by GPT-4 | ||
ShareGPT | Share your wildest ChatGPT conversations with one click | ||
Major LLMs + Data Availability | |||
500+ Best AI Tools | |||
Cohere Summarize Beta | Introducing Cohere Summarize Beta: A New Endpoint for Text Summarization | ||
chatgpt-wrapper | 3,659 | 18 days ago | ChatGPT Wrapper is an open-source unofficial Python API and CLI that lets you interact with ChatGPT |
Open-evals | 19 | over 1 year ago | A framework extend openai's for different language model |
Cursor | Write, edit, and chat about your code with a powerful AI | ||
AutoGPT | 168,407 | 4 days ago | an experimental open-source application showcasing the capabilities of the GPT-4 language model |
OpenAGI | 1,963 | 3 months ago | When LLM Meets Domain Experts |
EasyEdit | 1,931 | 6 days ago | An easy-to-use framework to edit large language models |
chatgpt-shroud | 9 | over 1 year ago | A Chrome extension for OpenAI's ChatGPT, enhancing user privacy by enabling easy hiding and unhiding of chat history. Ideal for privacy during screen shares |