Dela via


NGramTokenizer Class

Definition

Tokenizes the input into n-grams of the given size(s). This tokenizer is implemented using Apache Lucene. http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/ngram/NGramTokenizer.html

[Newtonsoft.Json.JsonObject("#Microsoft.Azure.Search.NGramTokenizer")]
public class NGramTokenizer : Microsoft.Azure.Search.Models.Tokenizer
[<Newtonsoft.Json.JsonObject("#Microsoft.Azure.Search.NGramTokenizer")>]
type NGramTokenizer = class
    inherit Tokenizer
Public Class NGramTokenizer
Inherits Tokenizer
Inheritance
NGramTokenizer
Attributes
Newtonsoft.Json.JsonObjectAttribute

Constructors

NGramTokenizer()

Initializes a new instance of the NGramTokenizer class.

NGramTokenizer(String, Nullable<Int32>, Nullable<Int32>, IList<TokenCharacterKind>)

Initializes a new instance of the NGramTokenizer class.

Properties

MaxGram

Gets or sets the maximum n-gram length. Default is 2. Maximum is 300.

MinGram

Gets or sets the minimum n-gram length. Default is 1. Maximum is 300. Must be less than the value of maxGram.

Name

Gets or sets the name of the tokenizer. It must only contain letters, digits, spaces, dashes or underscores, can only start and end with alphanumeric characters, and is limited to 128 characters.

(Inherited from Tokenizer)
TokenChars

Gets or sets character classes to keep in the tokens.

Methods

Validate()

Validate the object.

Applies to