567-labs / fastllm

A collection of LLM services you can self host via docker or modal labs to support your applications development
MIT License
182 stars 23 forks source link

Created a minimal reproduction #48

Closed ivanleomk closed 6 months ago

ivanleomk commented 6 months ago

This updates the optimize_plain.py script into a simpler standalone reproduction that we can use to run hyper-parameter sweeps. I updated the parameters inside the script to match that of the data that you gave Charles. Using n_trials=1 generates around 100+ unique configurations which is close to the number of trials inside the .csv that you gave to him.

Changes


:rocket: This description was created by Ellipsis for commit 9c40ffadf4e8cea15aad9957f01b9453b6ad36fc

Summary:

This PR simplifies the optimize_plain.py script for hyper-parameter sweeps, adds a deduplication step, converts the data class to a frozen data class, and adds a method to download the dataset from Quora.

Key points:


Generated with :heart: by ellipsis.dev