enable 3 char words to finegrind tokenize (#2210)

### What problem does this PR solve?


### Type of change


- [x] Performance Improvement
This commit is contained in:
Kevin Hu 2024-09-03 13:37:32 +08:00 committed by GitHub
parent 21179a9be9
commit 6d232f1bdb
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

View File

@ -83,7 +83,7 @@ class EsQueryer:
), tks
def need_fine_grained_tokenize(tk):
if len(tk) < 4:
if len(tk) < 3:
return False
if re.match(r"[0-9a-z\.\+#_\*-]+$", tk):
return False