Open elisalimli opened 6 months ago
Since we use gpt-3.5-turbo-16k model for summarising. If we try to summarise a doc is relatively long. It will not be able to do it.
https://github.com/superagent-ai/super-rag/blob/main/utils/summarise.py#L33
Example file: https://github.com/datasciencedojo/datasets/blob/master/titanic.csv
We will have to split the summaries into max 16k tokens with tiktokenizer or use a model with larger context.
Since we use gpt-3.5-turbo-16k model for summarising. If we try to summarise a doc is relatively long. It will not be able to do it.
https://github.com/superagent-ai/super-rag/blob/main/utils/summarise.py#L33
Example file: https://github.com/datasciencedojo/datasets/blob/master/titanic.csv