use this:

analyzer a1{
tokenizer = StandardTokenizer // split words
token filter = LowerCaseFilter
token filter = NGramFilter(minGramSize = “1”, maxGramSize = “50”) //index n-grams “t” “e” s" “t” “te” “es” "st “tes” “est” “test”
}

doesnt work:

analyzer a1{
tokenizer = PatternTokenizer(pattern=“([a-z])”, group=“1”)
token filter = LowerCaseFilter
}

analyzer a1{
tokenizer = PatternTokenizer(pattern=“([a-z]{2})”, group=“1”)
token filter = LowerCaseFilter
}

analyzer a1{
tokenizer = PatternTokenizer(pattern=“(.)”, group=“1”)
token filter = LowerCaseFilter
}

Submitted by Danny Groenewegen on 17 March 2022 at 14:38

Log in to post comments