Is token normalization implemented in Stanford NLP? -


i trying normalize tokens (potentially merging them if needed) before running regexner annotator on them.

is there implemented in stanford corenlp or in stanford nlp in general?

if not, what's best way implement it? writing custom annotator in corenlp?

there options token normalization. apply -options flag comma separated list containing options want.

this described in more detail on link:

http://nlp.stanford.edu/software/tokenizer.shtml

near bottom there section options shows list of possibilities.

are there other normalizations interested in not on list?


Comments

Popular posts from this blog

java - Andrioid studio start fail: Fatal error initializing 'null' -

android - Gradle sync Error:Configuration with name 'default' not found -

StringGrid issue in Delphi XE8 firemonkey mobile app -