Options
All
  • Public
  • Public/Protected
  • All
Menu

Enumeration KnownTokenFilterNames

Package version

Defines values for TokenFilterName.

readonly

Index

Enumeration members

Apostrophe

Apostrophe: = "apostrophe"

Strips all characters after an apostrophe (including the apostrophe itself). See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/tr/ApostropheFilter.html

ArabicNormalization

ArabicNormalization: = "arabic_normalization"

A token filter that applies the Arabic normalizer to normalize the orthography. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/ar/ArabicNormalizationFilter.html

AsciiFolding

AsciiFolding: = "asciifolding"

Converts alphabetic, numeric, and symbolic Unicode characters which are not in the first 127 ASCII characters (the "Basic Latin" Unicode block) into their ASCII equivalents, if such equivalents exist. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/miscellaneous/ASCIIFoldingFilter.html

CjkBigram

CjkBigram: = "cjk_bigram"

Forms bigrams of CJK terms that are generated from StandardTokenizer. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/cjk/CJKBigramFilter.html

CjkWidth

CjkWidth: = "cjk_width"

Normalizes CJK width differences. Folds fullwidth ASCII variants into the equivalent basic Latin, and half-width Katakana variants into the equivalent Kana. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/cjk/CJKWidthFilter.html

Classic

Classic: = "classic"

CommonGram

CommonGram: = "common_grams"

Construct bigrams for frequently occurring terms while indexing. Single terms are still indexed too, with bigrams overlaid. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/commongrams/CommonGramsFilter.html

EdgeNGram

EdgeNGram: = "edgeNGram_v2"

Generates n-grams of the given size(s) starting from the front or the back of an input token. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/ngram/EdgeNGramTokenFilter.html

Elision

Elision: = "elision"

Removes elisions. For example, "l'avion" (the plane) will be converted to "avion" (plane). See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/util/ElisionFilter.html

GermanNormalization

GermanNormalization: = "german_normalization"

Normalizes German characters according to the heuristics of the German2 snowball algorithm. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/de/GermanNormalizationFilter.html

HindiNormalization

HindiNormalization: = "hindi_normalization"

Normalizes text in Hindi to remove some differences in spelling variations. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/hi/HindiNormalizationFilter.html

IndicNormalization

IndicNormalization: = "indic_normalization"

KStem

KStem: = "kstem"

KeywordRepeat

KeywordRepeat: = "keyword_repeat"

Emits each incoming token twice, once as keyword and once as non-keyword. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/miscellaneous/KeywordRepeatFilter.html

Length

Length: = "length"

Limit

Limit: = "limit"

Lowercase

Lowercase: = "lowercase"

NGram

NGram: = "nGram_v2"

PersianNormalization

PersianNormalization: = "persian_normalization"

Phonetic

Phonetic: = "phonetic"

PorterStem

PorterStem: = "porter_stem"

Uses the Porter stemming algorithm to transform the token stream. See http://tartarus.org/~martin/PorterStemmer

Reverse

Reverse: = "reverse"

ScandinavianFoldingNormalization

ScandinavianFoldingNormalization: = "scandinavian_folding"

Folds Scandinavian characters åÅäæÄÆ->a and öÖøØ->o. It also discriminates against use of double vowels aa, ae, ao, oe and oo, leaving just the first one. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/miscellaneous/ScandinavianFoldingFilter.html

ScandinavianNormalization

ScandinavianNormalization: = "scandinavian_normalization"

Shingle

Shingle: = "shingle"

Snowball

Snowball: = "snowball"

SoraniNormalization

SoraniNormalization: = "sorani_normalization"

Stemmer

Stemmer: = "stemmer"

Stopwords

Stopwords: = "stopwords"

Trim

Trim: = "trim"

Truncate

Truncate: = "truncate"

Unique

Unique: = "unique"

Uppercase

Uppercase: = "uppercase"

WordDelimiter

WordDelimiter: = "word_delimiter"

Splits words into subwords and performs optional transformations on subword groups.

Generated using TypeDoc