Open szzhh opened 2 weeks ago
Thanks for the question! We indeed observed the variance in downstream classification accuracy, but not as large as shown in your figure. For each run, we use validation data to select the best epoch for downstream classification. We report the average accuracy over 3 runs.
Thank you for your reply above, I still have some questions I would like to consult about the paper 'Differentiated Private Synthetic Data via Foundation Model APIs 2: Text'.
RANDOM_API
, is the number of generated texts for each category random, without following the ratio of the two labels combined to form a new category?Thank you for your patience in reading my question. I am very much looking forward to your reply. Thank you very much!
Thanks for the great questions!
Feel free to ask if you have more questions!
Use Random_API combined with chatgpt3.5 to generate a set of data, then use Roberta_base to perform three downstream trainings and make predictions on the test set. For the business category, the results are stable, but for the rating category, the results of each run vary greatly, as shown in the figure. Do you encounter this problem during the experiment?