Open bwnjnOEI opened 3 weeks ago
Hey there, @bwnjnOEI! I'm here to help you with any bugs, questions, or contributions you might have. Let's dive into the issue you've raised.
In the Settings
class, the llm
parameter is used as the default language model for generating answers, which suggests it is likely used for both the ask LLM
and answer LLM
functionalities. This means that the llm
parameter could be serving as a model reuse for both asking questions and generating answers, rather than requiring separate models for each task [1].
In the example code snippet you provided, it appears that three LLMs are explicitly set up: agent_llm
, summary_llm
, and llm
. The llm
parameter is likely used for both asking and answering, which aligns with the idea of model reuse for ask LLM
and answer LLM
.
To continue talking to Dosu, mention @dosu.
setting
? In your paper PaperQA: Retrieval-Augmented Generative Agent for Scientific Research, this paper seems to have employing four LLM instances,agent LLM
,summary LLM
,ask LLM
andanswer LLM
. TheAgent LLM
adjust and iterate perform steps in response to each question, ensuring more precise and relevant answers; Ingather evidence
module, use thesummary LLM
summarize chunks and evaluate relevance score on retrieved chunk; Inanswer question
module, use theask LLM
to extract prior knowledge about the input query and useanswer LLM
for output the final answer. On the other side, in the example code snippet,imply you just need setup three LLMs:
agent_llm
andsummary_llm
started up just as the paper claimed, however,llm
? Is itask LLM
oranswer LLM
? Or for model reuse ofask LLM
andanswer LLM
(where only one model is used for both)? Is there anything I missed?summary_llm
to use a local 13B model while calling OpenAI or Anthropic's API for theagent_llm
?