Closed amritsingh183 closed 6 months ago
Hi here @amritsingh183! Thanks for opening the issue, indeed we're already working on this as well as aligning the supported params for other LLM providers too, I'll link the PR here once it's created so that you can use distilabel
from that branch until v1.1.0 is released!
Thanks @alvarobartt !!
Hi here @amritsingh183, the PR is still a draft but you can use it for n_ctx
with no issues now! Install it from the branch as pip install git+https://github.com/argilla-io/distilabel.git@align-llm-params
👍🏻
Also expect it to be released in ~2 weeks, follow the open roadmap to stay tuned of all the features, fixes and improvements that will come for distilabel
v1.1.0
Indeed this has just been merged into develop
, so feel free to install it from develop
instead 👍🏻
I tried the develop branch and it works... Thanks !! :-)
Is your feature request related to a problem? Please describe. I appreciate the work distilabel is doing and making it easier for the community to produce high quality datasets.Thank you!
There is a problem I have faced and a potential solution is outlined in this feature request.
Consider the following code where a smaller LLM is used for response generation and larger LLM is used for feedback
when this is run the following error pops up
Describe the solution you'd like Allow
n_ctx
in https://github.com/argilla-io/distilabel/blob/9f38b4931398f626e07cbe2a83ef393de661f428/src/distilabel/llms/llamacpp.py#L72 so that we have the ability to get thisBy adding
n_ctx
to https://github.com/argilla-io/distilabel/blob/9f38b4931398f626e07cbe2a83ef393de661f428/src/distilabel/llms/llamacpp.py#L76 the code works perfectlyDescribe alternatives you've considered no other option available
Additional context Here is some more data from the trace