Previously, we were passing in as many examples as we could within the constraints of context length, but it still wasn't good enough, which is why I tried splitting the intent determination up into two parts: first - is it a display query, and second - is it a CV or Docs query..
But that still wasn't good enough.
To overcome this, I've refactored so that we are doing similar example selection for the query intent examples we feed in to the intent classification model. I added a bunch more examples based on the FO docs, and had ChatGPT come up with some CV/ML examples as well.
The query_intent_examples.csv file only contains examples for documentation, computer vision, and invalid, and then we pull in and reuse our viewstage examples, labeling them with intent display.
Anecdotally, this results in much better intent classification. Plus, it cuts down on the potential number of distinct calls to the LLM.
Previously, we were passing in as many examples as we could within the constraints of context length, but it still wasn't good enough, which is why I tried splitting the intent determination up into two parts: first - is it a display query, and second - is it a CV or Docs query..
But that still wasn't good enough.
To overcome this, I've refactored so that we are doing similar example selection for the query intent examples we feed in to the intent classification model. I added a bunch more examples based on the FO docs, and had ChatGPT come up with some CV/ML examples as well.
The
query_intent_examples.csv
file only contains examples fordocumentation
,computer vision
, andinvalid
, and then we pull in and reuse our viewstage examples, labeling them with intentdisplay
.Anecdotally, this results in much better intent classification. Plus, it cuts down on the potential number of distinct calls to the LLM.