-
Notifications
You must be signed in to change notification settings - Fork 86
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
jieba插件对包含空格的字符串分词后,包含了值为空格的token #17
Comments
在尝试了使用trim和stop filter,都没办法过滤掉空格时, |
同义词那边回答里面我用了whitespace tokenizer ,所以空格都被过滤掉了 |
我是用
|
我按这个来,可以了。感谢 |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
使用jieba插件分词,对包含空格的字符串分词,会包含值为空格的token,search和index模式都是如此,比如:
那这样,如果用户搜索内容包括空格时,就有可能影响搜索结果了,因为搜索分词时包含空格,但是es索引的内容可能不包含空格。
The text was updated successfully, but these errors were encountered: