ProQA
QA Indexer
A resource-efficient method for pretraining dense corpus indexes for open-domain QA and IR.
Progressively Pretrained Dense Corpus Index for Open-Domain QA and Information Retrieval
43 stars
2 watching
2 forks
Language: Python
last commit: over 1 year ago
Linked from 1 awesome list
information-retrievalnatural-language-processingpytorchquestion-answering
Related projects:
Repository | Description | Stars |
---|---|---|
| A semantic indexer for Python code, enabling batch processing and interprocedural type inference to build indexes for code browsers and search engines. | 1,395 |
| A Python-based question answering system built on top of Notion's database and OpenAI's API for natural language processing. | 2,139 |
| Tools and codebase for training neural question answering models on multiple paragraphs of text data | 435 |
| A collection of Python-related questions and answers from Stack Overflow, with some answers translated into Chinese. | 856 |
| A solver for optimizing quadratic programs in deep learning and other fields. | 691 |
| Implementing reading comprehension from Wikipedia questions to answer open-domain queries using PyTorch and SQuAD dataset | 401 |
| Improves search quality across multiple search engines and platforms. | 285 |
| This project provides code for training image question answering models using stacked attention networks and convolutional neural networks. | 108 |
| A Python library for efficiently indexing and searching large datasets. | 31 |
| Python bindings for Coq's interactive proof assistant | 50 |
| An insurance industry conversation corpus with pre-processed data for natural language processing and question answering tasks. | 1,019 |
| An application for quality assurance of scanned book or newspaper images with detailed annotations. | 22 |
| An API for evaluating the quality of text prompts used in Large Language Models (LLMs) based on perplexity estimation | 43 |
| An eXplainability toolbox for machine learning that enables data analysis and model evaluation to mitigate biases and improve performance | 1,135 |
| A framework for evaluating and monitoring the quality of large language model outputs in Retrieval Augmented Generation applications. | 271 |