teamtnt / tntsearch

A fully featured full text search engine written in PHP
https://tnt.studio/solving-the-search-problem-with-laravel-and-tntsearch
MIT License
3.1k stars 291 forks source link

Scout: Custom tokenizer indexing properly to allow dashes and periods, but searching on dashes does not work #290

Open bretvanhorn opened 1 year ago

bretvanhorn commented 1 year ago

Hi, I have created a custom tokenizer that allows dashes, plus signs, and periods in the indexed keywords. I've verified it is correctly indexing such terms in the index:

Screenshot 2023-06-26 at 10 56 30 AM

However, when I search, it does not return any of the indexed items that should match:

Screenshot 2023-06-26 at 10 56 39 AM

Here is my config:

    'tntsearch' => [
        'storage'  => storage_path(), //place where the index files will be stored
        'fuzziness' => env('TNTSEARCH_FUZZINESS', false),
        'fuzzy' => [
            'prefix_length' =>  env('TNTSEARCH_FUZZY_LEN', 2),
            'max_expansions' =>  env('TNTSEARCH_FUZZY_EXPANSIONS', 50),
            'distance' =>  env('TNTSEARCH_FUZZY_DISTANCE', 2),
            'no_limit' =>  env('TNTSEARCH_FUZZY_NO_LIMIT', false)
        ],
        'asYouType' => false,
        'searchBoolean' => env('TNTSEARCH_BOOLEAN', false),
        'maxDocs' => env('TNTSEARCH_MAX_DOCS', 500),
        'tokenizer' => \App\Http\Classes\ItemTokenizer::class
    ],

And the .env vars referenced:

SCOUT_DRIVER=tntsearch
SCOUT_QUEUE=false
TNTSEARCH_FUZZINESS=true
TNTSEARCH_BOOLEAN=false
TNTSEARCH_MAX_DOCS=2500
TNTSEARCH_FUZZY_LEN=2
TNTSEARCH_FUZZY_EXPANSIONS=500
TNTSEARCH_FUZZY_DISTANCE=2
TNTSEARCH_FUZZY_NO_LIMIT=false

Here is my tokenizer:

    namespace App\Http\Classes;

    use TeamTNT\TNTSearch\Support\AbstractTokenizer;
    use TeamTNT\TNTSearch\Support\TokenizerInterface;

    class ItemTokenizer extends AbstractTokenizer implements TokenizerInterface {

        static protected $pattern = '/[^\p{L}\p{N}\.\+-]+/u';

        public function tokenize($text, $stopwords = []) {
            return preg_split($this->getPattern(), strtolower($text), -1, PREG_SPLIT_NO_EMPTY);
        }
    }

I am not sure if this issue lies with the Scout plugin or the core engine, so please let me know if I need to move this to the Scout plugin issues and/or any other info I can provide.

nticaric commented 1 year ago

Thank you for bringing this to our attention. To help us understand the issue better, could you please try implementing a diagnostic step in your code?

You can use the dd() function within your custom tokenizer during a search operation. This way we'll see if it hits the correct tokenizer or the default one


public function tokenize($text, $stopwords = []) {
    $return = preg_split($this->getPattern(), strtolower($text), -1, PREG_SPLIT_NO_EMPTY);
    dd($return);    
}
nticaric commented 1 year ago

And while testing this, please set the fuzziness to false

TNTSEARCH_FUZZINESS=false
bretvanhorn commented 1 year ago

@nticaric Thanks for the quick reply! So, I tried this and interestingly enough, there is no debug output. This search is being done via an API call, but when I go to the API server and view the request for the search in Debugbar, there is no dd output. I am not sure what I am doing given this revealation, but any suggestions would be welcome!

nticaric commented 1 year ago

Can you query the info table of the index? It could be that the original index was build with the default tokenizer

bretvanhorn commented 1 year ago

@nticaric, here you go:

Screenshot 2023-06-26 at 2 09 07 PM
bretvanhorn commented 1 year ago

Ok, I had forgotten to set the return as a var, so it was returning before the dd was called. Now it appears to be breaking the request, which tells me it is using the custom tokenizer.

If it helps, it almost looks as though it is struggling with numerical content at the beginning of the keyword. For example, sx-70 seems to return relevant results, but 70-200 returns items with either 70 or 200 in the name and a dash in the name outside of that, but prioritizes 200.

And I remember that the import seems to have indexed just simply "-" as a keyword, and I wonder if that is the issue. I am going to try to delete that keyword from the keywords and see what that does.

nticaric commented 1 year ago

Are you sure fuzziness is turned off?

bretvanhorn commented 1 year ago

Are you sure fuzziness is turned off?

Yep, confirmed. I realized my regexp is allowing spaces and dashes with spaces around them to be indexed as keywords. I am slow with Regexp, so I am trying to remedy that now.

bretvanhorn commented 1 year ago

Ok @nticaric here is the regex I am currently using:

static protected $pattern = '/[^\p{L}\p{N}\.\+-](?!\s-\s)+/u';

I am still seeing the behavior where 70-200mm does not return relevant results (items with "200" and "70" show up, but nothing with "70-200"), but sx-70 returns relevant results. Any thoughts or guidance is welcome. Again, I suck at regex, so perhaps there is a better way to say:

Do not allow the following characters to be treated as stop words:

Again indexing seems to allow phrases like 70-200mm into the wordlist table, but searching for them does not yield expected results.

Screenshot 2023-06-26 at 3 20 39 PM

Thanks again for your help.