例如,我将一只股票名 “深深宝A”加入了用户词典,分词之后仍然是如下结果,
{
"tokens": [
{
"token": "深深",
"start_offset": 0,
"end_offset": 2,
"type": "CN_WORD",
"position": 0
},
{
"token": "宝",
"start_offset": 2,
"end_offset": 3,
"type": "CN_CHAR",
"position": 1
},
{
"token": "a",
"start_offset": 3,
"end_offset": 4,
"type": "ENGLISH",
"position": 2
}
]
}
有没有遇到这种情况的,我初步猜测是有英文a的原因,不知道如何解决
{
"tokens": [
{
"token": "深深",
"start_offset": 0,
"end_offset": 2,
"type": "CN_WORD",
"position": 0
},
{
"token": "宝",
"start_offset": 2,
"end_offset": 3,
"type": "CN_CHAR",
"position": 1
},
{
"token": "a",
"start_offset": 3,
"end_offset": 4,
"type": "ENGLISH",
"position": 2
}
]
}
有没有遇到这种情况的,我初步猜测是有英文a的原因,不知道如何解决
3 个回复
jiangtao
赞同来自:
我将其添加到词库中,分词结果如下:
laoyang360 - 《一本书讲透Elasticsearch》作者,Elastic认证工程师 [死磕Elasitcsearch]知识星球地址:http://t.cn/RmwM3N9;微信公众号:铭毅天下; 博客:https://elastic.blog.csdn.net
赞同来自:
phoenix
赞同来自: