Closed Bri1987 closed 4 months ago
Hi😄! Thank you for your report !!! The FunSearch pipeline does not need the '_additional_prompt'. More accurately, FunSearch is designed to work with ''code completion style models'' (e.g., StarCoder, CodeLlama, ...), for which additional prompts and instructions are not really required. For other "instruct-based models" such as DeepSeek-coder, LLama, GPT-3.5..., it is not sure if adding a prompt based on the original prompt can help better/quicker convergence. You may test it !!! 😁
ok, thanks you a lot !
Thank you for your great work! But I noticed that in the LocalLLM class of the "funsearch_bin_packing_local_llm.py" file, the _additional_prompt is initialized in the init method. However, it seems that it is not passed to the LLM-server through the data. Is it a bug or I don't understand your work entirely ?