MicrosoftLanguageStemmingTokenizer Class
Definition
Important
Some information relates to prerelease product that may be substantially modified before it’s released. Microsoft makes no warranties, express or implied, with respect to the information provided here.
Divides text using language-specific rules and reduces words to their base forms.
[Newtonsoft.Json.JsonObject("#Microsoft.Azure.Search.MicrosoftLanguageStemmingTokenizer")]
public class MicrosoftLanguageStemmingTokenizer : Microsoft.Azure.Search.Models.Tokenizer
[<Newtonsoft.Json.JsonObject("#Microsoft.Azure.Search.MicrosoftLanguageStemmingTokenizer")>]
type MicrosoftLanguageStemmingTokenizer = class
inherit Tokenizer
Public Class MicrosoftLanguageStemmingTokenizer
Inherits Tokenizer
- Inheritance
- Attributes
-
Newtonsoft.Json.JsonObjectAttribute
Constructors
MicrosoftLanguageStemmingTokenizer() |
Initializes a new instance of the MicrosoftLanguageStemmingTokenizer class. |
MicrosoftLanguageStemmingTokenizer(String, Nullable<Int32>, Nullable<Boolean>, Nullable<MicrosoftStemmingTokenizerLanguage>) |
Initializes a new instance of the MicrosoftLanguageStemmingTokenizer class. |
Properties
IsSearchTokenizer |
Gets or sets a value indicating how the tokenizer is used. Set to true if used as the search tokenizer, set to false if used as the indexing tokenizer. Default is false. |
Language |
Gets or sets the language to use. The default is English. Possible values include: 'arabic', 'bangla', 'bulgarian', 'catalan', 'croatian', 'czech', 'danish', 'dutch', 'english', 'estonian', 'finnish', 'french', 'german', 'greek', 'gujarati', 'hebrew', 'hindi', 'hungarian', 'icelandic', 'indonesian', 'italian', 'kannada', 'latvian', 'lithuanian', 'malay', 'malayalam', 'marathi', 'norwegianBokmaal', 'polish', 'portuguese', 'portugueseBrazilian', 'punjabi', 'romanian', 'russian', 'serbianCyrillic', 'serbianLatin', 'slovak', 'slovenian', 'spanish', 'swedish', 'tamil', 'telugu', 'turkish', 'ukrainian', 'urdu' |
MaxTokenLength |
Gets or sets the maximum token length. Tokens longer than the maximum length are split. Maximum token length that can be used is 300 characters. Tokens longer than 300 characters are first split into tokens of length 300 and then each of those tokens is split based on the max token length set. Default is 255. |
Name |
Gets or sets the name of the tokenizer. It must only contain letters, digits, spaces, dashes or underscores, can only start and end with alphanumeric characters, and is limited to 128 characters. (Inherited from Tokenizer) |
Methods
Validate() |
Validate the object. |
Applies to
意見反應
https://aka.ms/ContentUserFeedback。
即將登場:在 2024 年,我們將逐步淘汰 GitHub 問題作為內容的意見反應機制,並將它取代為新的意見反應系統。 如需詳細資訊,請參閱:提交並檢視相關的意見反應