opik

LLM tester

An end-to-end platform for evaluating and testing large language models.

Open-source end-to-end LLM Development Platform

GitHub

2k stars
29 watching
131 forks
Language: Java
last commit: 6 days ago
Linked from 9 awesome lists


Backlinks from these awesome lists:

Related projects:

Repository Description Stars
poyro/poyro An extension of Vitest for testing LLM applications using local language models 30
innogames/ltc A tool for managing load tests and analyzing performance results 198
lcm-proj/lcm A set of libraries and tools for efficient message passing and data marshalling in real-time systems. 1,003
sunlemuria/opengptandbeyond An effort to develop and compare large language models beyond OpenGPT 105
johnsnowlabs/langtest A tool for testing and evaluating large language models with a focus on AI safety and model assessment. 501
luogen1996/lavin An open-source implementation of a vision-language instructed large language model 508
norman/telescope A test library for Lua that supports declarative testing with nested contexts and code coverage reports. 161
qcri/llmebench A benchmarking framework for large language models 80
openolat/openolat A web-based e-learning platform with features like assessment, content management, and learning resources, built using Java. 332
llm-ui-kit/llm-ui A React library designed to work with Large Language Models (LLMs) by providing features such as syntax removal, custom component addition, and rendering at a native frame rate. 403
ailab-cvc/seed-bench A benchmark for evaluating large language models' ability to process multimodal input 315
melih-unsal/demogpt A comprehensive toolset for building Large Language Model (LLM) based applications 1,710
talkdai/dialog An application framework to simplify the deployment and testing of large language models (LLMs) for natural language processing tasks. 377
davidmigloz/langchain_dart Provides a set of tools and components to simplify the integration of Large Language Models into Dart/Flutter applications 425
mlabonne/llm-autoeval A tool to automate the evaluation of large language models in Google Colab using various benchmarks and custom parameters. 558