doc2term

A fast NLP tokenizer that detects tokens and remove duplications and punctuations


Keywords
tokenizer, NLP, punctuation, standarization, duplicate-detection, text-processing, text-tokenizing, doc2term
License
Apache-2.0
Install
pip install doc2term==0.1

Documentation

doc2term

Build Status pypi license

A fast NLP tokenizer that detects sentences, words, numbers, urls, hostnames, emails, filenames, and phone numbers. Tokenize integrates and standardize the documents, remove the punctuations and duplications.

Installation

pip install doc2term

Compilation

The installation requires to compile the original C code using gcc.

Usage

Example notebook: doc2term

Example

>>> import doc2term

>>> doc2term.doc2term_str("Actions speak louder than words. ... ")
"Actions speak louder than words ."
>>> doc2term.doc2term_str("You can't judge a book by its cover. ... from thoughtcatalog.com")
"You can't judge a book by its cover . from thoughtcatalog.com"