opik

LM testing platform

A platform for evaluating and testing large language models (LLMs) during development and production.

Open-source end-to-end LLM Development Platform

GitHub

3k stars
38 watching
158 forks
Language: Java
last commit: about 1 month ago
Linked from 9 awesome lists


Backlinks from these awesome lists:

Related projects:

Repository Description Stars
poyro/poyro An extension of Vitest for testing LLM applications using local language models 31
innogames/ltc A tool for managing load tests and analyzing performance results 200
lcm-proj/lcm A set of libraries and tools for efficient message passing and data marshalling in real-time systems. 1,011
sunlemuria/opengptandbeyond An effort to develop and compare large language models beyond OpenGPT 105
johnsnowlabs/langtest A tool for testing and evaluating large language models with a focus on AI safety and model assessment. 506
luogen1996/lavin An open-source implementation of a vision-language instructed large language model 513
norman/telescope A test library for Lua that supports declarative testing with nested contexts and code coverage reports. 161
qcri/llmebench A benchmarking framework for large language models 81
openolat/openolat A web-based e-learning platform with features like assessment, content management, and learning resources, built using Java. 337
llm-ui-kit/llm-ui A React library designed to work with Large Language Models (LLMs) by providing features such as syntax removal, custom component addition, and rendering at a native frame rate. 425
ailab-cvc/seed-bench A benchmark for evaluating large language models' ability to process multimodal input 322
melih-unsal/demogpt A comprehensive toolset for building Large Language Model (LLM) based applications 1,733
talkdai/dialog An application framework to simplify the deployment and testing of large language models (LLMs) for natural language processing tasks. 380
davidmigloz/langchain_dart Provides a set of tools and components to simplify the integration of Large Language Models into Dart/Flutter applications 441
mlabonne/llm-autoeval A tool to automate the evaluation of large language models in Google Colab using various benchmarks and custom parameters. 566