bleurt
NLG evaluation metric
An evaluation metric for Natural Language Generation based on transfer learning.
BLEURT is a metric for Natural Language Generation based on transfer learning.
705 stars
13 watching
85 forks
Language: Python
last commit: over 2 years ago
Linked from 1 awesome list
Related projects:
| Repository | Description | Stars |
|---|---|---|
| | A toolset for evaluating and comparing natural language generation models | 1,350 |
| | Provides implementations of various supervised machine learning evaluation metrics in multiple programming languages. | 1,632 |
| | A repository of papers and resources for evaluating large language models. | 1,450 |
| | A statistical natural language parser used to generate grammatically correct sentences from unstructured text input. | 227 |
| | A collection of research implementations and models for natural language generation | 694 |
| | Compiles bias evaluation datasets and provides access to original data sources for large language models | 115 |
| | Provides intermediate representations of data for NLG tasks like Discourse Ordering and Lexicalization | 69 |
| | A Python library to manipulate and transform indexable data | 49 |
| | A tool for calculating and analyzing BPEL metrics | 0 |
| | A Java API for generating natural language texts from syntactic forms | 810 |
| | A framework for efficient and optimized retrieval augmented generative pipelines using state-of-the-art LLMs and Information Retrieval. | 1,392 |
| | A library providing an implementation of various metrics for object segmentation and saliency detection in computer vision. | 150 |
| | A pre-trained language model designed for various NLP tasks, including dialogue generation, code completion, and retrieval. | 94 |
| | Provides benchmarking policies and datasets for offline reinforcement learning | 85 |