punkt-segmenter
Sentence tokenizer
A Ruby port of the NLTK algorithm to detect sentence boundaries in unstructured text
Ruby port of the NLTK Punkt sentence segmentation algorithm
92 stars
2 watching
10 forks
Language: Ruby
last commit: over 6 years ago
Linked from 1 awesome list
nlp-librarynltkpunkt-segmenterrubyruby-portrubynlpsentence-boundariessentence-tokenizertokenized-sentences
Related projects:
Repository | Description | Stars |
---|---|---|
| A Ruby port of a Japanese text tokenization algorithm | 21 |
| A rule-based sentence boundary detection gem that works across many languages | 559 |
| A Ruby library that tokenizes text into sentences using a Bayesian statistical model | 80 |
| A Ruby-based library for splitting written text into tokens for natural language processing tasks. | 46 |
| Breaks text into contiguous sequences of words or phrases | 12 |
| A Ruby library that tokenizes input and provides various statistical measures about the tokens | 159 |
| A gem for extracting words from text with customizable tokenization rules | 31 |
| A Python package for out-of-the-box sentence boundary detection using rule-based algorithms. | 821 |
| A fast and simple tokenizer for multiple languages | 28 |
| A Python wrapper around the Thai word segmentator LexTo, allowing developers to easily integrate it into their applications. | 1 |
| A tokeniser for natural language text that separates words from punctuation and supports basic preprocessing steps such as case changing | 66 |
| A multilingual tokenizer to split strings into tokens, handling various language and formatting nuances. | 90 |
| A Ruby wrapper around the Stuttgarter Tree Tagger for natural language processing tasks | 6 |
| A Ruby library that uses a simple rule-based approach to segment sentences into individual words or phrases. | 51 |
| A Ruby gem for pre-processing markdown files with file inclusion and formatting options. | 40 |