Is token normalization implemented in Stanford NLP? -


i trying normalize tokens (potentially merging them if needed) before running regexner annotator on them.

is there implemented in stanford corenlp or in stanford nlp in general?

if not, what's best way implement it? writing custom annotator in corenlp?

there options token normalization. apply -options flag comma separated list containing options want.

this described in more detail on link:

http://nlp.stanford.edu/software/tokenizer.shtml

near bottom there section options shows list of possibilities.

are there other normalizations interested in not on list?


Comments

Popular posts from this blog

android - Gradle sync Error:Configuration with name 'default' not found -

java - Andrioid studio start fail: Fatal error initializing 'null' -

html - jQuery UI Sortable - Remove placeholder after item is dropped -