tokenize_ptb {tokenizers} | R Documentation |
Penn Treebank Tokenizer
Description
This function implements the Penn Treebank word tokenizer.
Usage
tokenize_ptb(x, lowercase = FALSE, simplify = FALSE)
Arguments
x |
A character vector or a list of character vectors to be tokenized
into n-grams. If |
lowercase |
Should the tokens be made lower case? |
simplify |
|
Details
This tokenizer uses regular expressions to tokenize text similar to the tokenization used in the Penn Treebank. It assumes that text has already been split into sentences. The tokenizer does the following:
splits common English contractions, e.g.
don't
is tokenized intodo n't
andthey'll
is tokenized into ->they 'll
,handles punctuation characters as separate tokens,
splits commas and single quotes off from words, when they are followed by whitespace,
splits off periods that occur at the end of the sentence.
This function is a port of the Python NLTK version of the Penn Treebank Tokenizer.
Value
A list of character vectors containing the tokens, with one element
in the list for each element that was passed as input. If simplify =
TRUE
and only a single element was passed as input, then the output is a
character vector of tokens.
References
Examples
song <- list(paste0("How many roads must a man walk down\n",
"Before you call him a man?"),
paste0("How many seas must a white dove sail\n",
"Before she sleeps in the sand?\n"),
paste0("How many times must the cannonballs fly\n",
"Before they're forever banned?\n"),
"The answer, my friend, is blowin' in the wind.",
"The answer is blowin' in the wind.")
tokenize_ptb(song)
tokenize_ptb(c("Good muffins cost $3.88\nin New York. Please buy me\ntwo of them.",
"They'll save and invest more.",
"Hi, I can't say hello."))