Open bretvanhorn opened 1 year ago
Thank you for bringing this to our attention. To help us understand the issue better, could you please try implementing a diagnostic step in your code?
You can use the dd() function within your custom tokenizer during a search operation. This way we'll see if it hits the correct tokenizer or the default one
public function tokenize($text, $stopwords = []) {
$return = preg_split($this->getPattern(), strtolower($text), -1, PREG_SPLIT_NO_EMPTY);
dd($return);
}
And while testing this, please set the fuzziness to false
TNTSEARCH_FUZZINESS=false
@nticaric Thanks for the quick reply! So, I tried this and interestingly enough, there is no debug output. This search is being done via an API call, but when I go to the API server and view the request for the search in Debugbar, there is no dd output. I am not sure what I am doing given this revealation, but any suggestions would be welcome!
Can you query the info
table of the index? It could be that the original index was build with the default tokenizer
@nticaric, here you go:
Ok, I had forgotten to set the return as a var, so it was returning before the dd was called. Now it appears to be breaking the request, which tells me it is using the custom tokenizer.
If it helps, it almost looks as though it is struggling with numerical content at the beginning of the keyword. For example, sx-70 seems to return relevant results, but 70-200 returns items with either 70 or 200 in the name and a dash in the name outside of that, but prioritizes 200.
And I remember that the import seems to have indexed just simply "-" as a keyword, and I wonder if that is the issue. I am going to try to delete that keyword from the keywords and see what that does.
Are you sure fuzziness is turned off?
Are you sure fuzziness is turned off?
Yep, confirmed. I realized my regexp is allowing spaces and dashes with spaces around them to be indexed as keywords. I am slow with Regexp, so I am trying to remedy that now.
Ok @nticaric here is the regex I am currently using:
static protected $pattern = '/[^\p{L}\p{N}\.\+-](?!\s-\s)+/u';
I am still seeing the behavior where 70-200mm does not return relevant results (items with "200" and "70" show up, but nothing with "70-200"), but sx-70 returns relevant results. Any thoughts or guidance is welcome. Again, I suck at regex, so perhaps there is a better way to say:
Do not allow the following characters to be treated as stop words:
Again indexing seems to allow phrases like 70-200mm into the wordlist table, but searching for them does not yield expected results.
Thanks again for your help.
Hi, I have created a custom tokenizer that allows dashes, plus signs, and periods in the indexed keywords. I've verified it is correctly indexing such terms in the index:
However, when I search, it does not return any of the indexed items that should match:
Here is my config:
And the .env vars referenced:
Here is my tokenizer:
I am not sure if this issue lies with the Scout plugin or the core engine, so please let me know if I need to move this to the Scout plugin issues and/or any other info I can provide.