Token Filters #
Token filters accept a stream of tokens from a tokenizer and can modify tokens (eg lowercasing), delete tokens (eg remove stopwords) or add tokens (eg synonyms).
Elasticsearch has a number of built in token filters which can be used to build custom analyzers.
Standard Token Filter #
standard
Currently does nothing.
ASCII Folding Token Filter #
asciifolding
Converts alphabetic, numeric, and symbolic Unicode characters which are not in the first 127 ASCII characters (the “Basic Latin” Unicode block) into their ASCII equivalents, if one exists.
Length Token Filter #
length
Lowercase Token Filter #
lowercase
Uppercase Token Filter #
uppercase
NGram Token Filter #
nGram
Edge NGram Token Filter #
edgeNGram
Porter Stem Token Filter #
porter_stem
Shingle Token Filter #
shingle
Stop Token Filter #
stop
Word Delimiter Token Filter #
word_delimiter