标准分词器在哪些字符上定界?

What characters does the standard tokenizer delimit on?

我想知道哪些字符用于分隔 elasticsearch 标准分词器的字符串?

根据 documentation I believe this is the list of symbols/characters used for defining tokens: http://unicode.org/reports/tr29/#Default_Word_Boundaries