"token" : "x", "start_offset" : 2, "end_offset" : 3, "type" : "CN_CHAR", "position" : 1 }, { "token" : "x", "start_offset" : 3, "end_offset" : 4, "type" : "CN_CHAR", "position" : 2 }, { "token" : "真的", "start_offset" : 4, "end_offset" : 6, "...
standard tokenizer:以单词边界进行切分 standard token filter:什么都不做 lowercase token filter:将所有字母转换为小写 stop token filer(默认被禁用):移除停用词,比如a the it等等 11.2.2 修改分词器的设置 启用english停用词token filter PUT /my_index { "settings": { "analysis": { "analyzer": { "es_...