Open irthomasthomas opened 1 year ago
Quick notes. 1A: "Token wise but foolish" no tags. = Very low relevance. No answers about llm tokens. 1B: "Token wise but foolish. #llm #gpt #chatgpt" = Very high number of answers relevant to llms
2A: " wise but foolish" no tags = As expected, no relevance to llms 2B: "__wise but _____ foolish #llm #gpt #chatgpt" = Surprisingly, the inclusion of the tags did not improve the relevance of the answers.
Experiment 1
Prompt A: Token wise but _____ foolish.
Results for gpt-3.5-turbo-0613 temperature=1.
Prompt B: Token wise but _____ foolish. #llm #gpt #chatgpt
Results for gpt-3.5-turbo-0613 temperature=1.
Experiment 2
Prompt A: wise but foolish
Results for gpt-3.5-turbo-0613 temperature=1.
Prompt B: wise but foolish. #llm #gpt #chatgpt.
Results:
Summary:
Experiment 1:
Experiment 2:
Overall, the responses from GPT-3.5-turbo-0613 at a temperature of 1 varied in their specificity and creativity. Some responses provided unique combinations of wise and foolish, while others were more general in nature. The inclusion of keywords in the prompts seemed to influence the types of foolishness mentioned in the responses.