Chainlit / chainlit

Build Conversational AI in minutes ⚡️
https://docs.chainlit.io
Apache License 2.0
6.25k stars 799 forks source link

-w argument in the chainlit run command doesn't work yet #29

Closed ghost closed 1 year ago

ghost commented 1 year ago

As title, it's an honor to start developing chat based UIs swiftly through chainlit, it would be great if -w argument starts working in the windows in the upcoming updates.

willydouhard commented 1 year ago

Glad to hear Chainlit fits your needs! I still remember our conversation and I did not forget about this issue 😄. We will dig into it and hopefully fix it soon!

ghost commented 1 year ago

I am in no hurry and this issue has been opened as suggested while other one was closed.

willydouhard commented 1 year ago

Could you try to:

  1. install chainlit from the wheel provided in this PR https://github.com/Chainlit/chainlit/pull/40
  2. run chainlit hello

I would like to see if the rework fix those issues. Note that this version is not merged yet. The infos to install are at the end of the PR's description.

ghost commented 1 year ago

-w argument still doesn't work

willydouhard commented 1 year ago

Should be fixed in the latest version 0.3.0. Please note that it contains breaking changes. We prepared a migration guide to make it easy for everyone.

ghost commented 1 year ago

-w argument works! but there's a problem with async while running llms locallly

With @cl.langchain_factory(use_async=True)

the following error appears

Screenshot 2023-06-14 055114

willydouhard commented 1 year ago

This means the LLM you are using does not have an async implementation (at the langchain level, not chainlit). Whenever possible you should use the async implementation, but in that case you will have to fallback to the sync implementation as it is the only one available. To do that change your factory with @cl.langchain_factory(use_async=False)