Closed 2018211801 closed 1 year ago
I noticed that your test dataset has 11810 examples, how many examples can I put at least?
Please be aware that replicating our experiment may involve certain costs, considering the usage of the davinci model and the large number of examples (11810).
To answer your first question, please note that the predicted results will automatically be saved in the predicted_examples.jsonl
file within the output directory. If you run the experiment again with the same output path, the model will skip the samples whose responses have already been generated and will only append the results for the remaining samples.
Regarding your second question, you have the option to modify the list of tasks in the data/splits/default/test_tasks.txt
file to include only a single task, which will then result in 100 instances in total. If you wish to further minimize the number of instances, you can adjust the max_num_instances_per_eval_task
parameter in your run_ICIL.sh
file to the desired number.
Now, To provide you with a rough estimate of the cost, since we are forcing a 2048 max input length, and the cost is $0.02 per 1K token, running a single task with 100 instances will result in $4. In a similar manner, fully replicating our results (11810) will result in $400~$500.
Thank you very much for your patient answer. I was too careless. I have decided to try the open source model GPT-J 6B first.
I just wanted to reproduce your run_ICIL part of the experiment, and it turned out that it cost 28¥ before the experiment was finished!!!Is there a way to save it?