llama_deploy
Deployment framework
An async-first framework for deploying and scaling agentic multi-service systems
Deploy your agentic worfklows to production
2k stars
24 watching
198 forks
Language: Python
last commit: 11 months ago
Linked from 1 awesome list
agentsdeploymentframeworkllamaindexllmmulti-agents
Related projects:
| Repository | Description | Stars |
|---|---|---|
| | A data framework for integrating large language models into applications with custom data | 1,997 |
| | An application framework to simplify the deployment and testing of large language models (LLMs) for natural language processing tasks. | 380 |
| | An implementation of LLaMA2 in Docker, allowing users to quickly deploy and run the model on their local machine. | 537 |
| | Provides tools and frameworks for deploying machine learning models in production environments | 73 |
| | A Common Lisp port of a Large Language Model (LLM) implementation | 36 |
| | A tool for easy interaction with Large Language Models (LLMs) to execute structured function calls and generate structured output. | 505 |
| | An end-to-end platform for building and deploying large language model applications | 1,624 |
| | Provides a REST API for deploying and managing LangChain runnables and chains | 1,970 |
| | A framework and benchmark for training and evaluating multi-modal large language models, enabling the development of AI agents capable of seamless interaction between humans and machines. | 305 |
| | A collection of pre-configured environments for deploying various Google Cloud-based applications | 730 |
| | A lightweight framework for building agent-based applications using LLMs and transformer architectures | 1,924 |
| | A library for creating high-performance environments for training large numbers of competing agents in multi-agent scenarios | 240 |
| | A Slack bot that automates the deployment process of web applications to various environments. | 38 |
| | Provides an architectural approach and reference implementation to deploy workload platforms on Azure at scale | 46 |
| | A high-performance LLM written in Python/Jax for training and inference on Google Cloud TPUs and GPUs. | 1,557 |