Options
All
  • Public
  • Public/Protected
  • All
Menu

Interface StandardTokenizer

Package version

Breaks text following the Unicode Text Segmentation rules. This tokenizer is implemented using Apache Lucene.

Hierarchy

  • StandardTokenizer

Index

Properties

Optional maxTokenLength

maxTokenLength: undefined | number

The maximum token length. Default is 255. Tokens longer than the maximum length are split. Default value: 255.

name

name: string

The name of the tokenizer. It must only contain letters, digits, spaces, dashes or underscores, can only start and end with alphanumeric characters, and is limited to 128 characters.

odatatype

odatatype: "#Microsoft.Azure.Search.StandardTokenizer"

Polymorphic Discriminator

Generated using TypeDoc