标准分词器在哪些字符上定界?
What characters does the standard tokenizer delimit on?
我想知道哪些字符用于分隔 elasticsearch 标准分词器的字符串?
根据 documentation I believe this is the list of symbols/characters used for defining tokens: http://unicode.org/reports/tr29/#Default_Word_Boundaries
我想知道哪些字符用于分隔 elasticsearch 标准分词器的字符串?
根据 documentation I believe this is the list of symbols/characters used for defining tokens: http://unicode.org/reports/tr29/#Default_Word_Boundaries