RuohoRecords / Riddim-Archive

Archive of creative-commons Dub, Reggae, Dancehall and other Soundsystem riddims
13 stars 5 forks source link

Brainstorm ISMIR tasks #3

Open ruohoruotsi opened 5 years ago

ruohoruotsi commented 5 years ago

Brainstorm ISMIR 2020 tasks that would be enlightened, liberated, illuminated with a proper MIDI dataset of Caribbean riddims (some with melodies and skank chord progressions)

ruohoruotsi commented 5 years ago

This is interesting:

Cortexelus commented 4 years ago

Generation -- could try taking a pre-trained Transformer (trained on a large multi-genre midi collection) then fine-tune it on this dataset [like the LakhNES training method]

ruohoruotsi commented 4 years ago

Hey CJ, that's great idea ... so to summarize (1) preprocess data into trainable format (2) finetune on something like the Transformer-XL checkpoints (3) generate more basslines. What did I miss?

Cortexelus commented 4 years ago

That sounds about right. Convert your dataset into the format the pretrained model used. Finetune train (swap in this new dataset and continue training). Generate tons of output examples as you train.

Not sure if you've done other generative Transformer finetuning experiments, but I've done it on text with GPT2, as far as output quality, there's usually a sweet range. Too little iterations and it doesn't fit the aesthetic of your dataset yet. Too many iterations and it plagiarizes too much. The sweet range may last several thousand iterations. Every iteration you sample from has a different personality. If you only generate from your best/last iteration, the output won't be as diverse as if you were to generate from many different iterations in the sweet range.

ruohoruotsi commented 4 years ago

Thanks CJ, I appreciate the tips! Once I'm going on the dataset, and have some initial generated basslines. I'll share it w/ you, perhaps there's a vector of collaboration here.

I haven't fine-tuned generative Transformers. But I have built/iterated/refined lots of other audio/dsp/speech models mostly on audio-signals though.

Do you have any recommendations on MIDI python libs/frameworks that'd be useful in whipping the MIDI data into shape? I fear I may have to manually pull the "riddim"/bassline out of the songs that are multitrack, others are just riddims + drum. Also debating whether to keep the "skank" with the bassline 🙇 🤔

Cortexelus commented 4 years ago

cool! :D keep me in the loop

hmm try pretty-midi http://craffel.github.io/pretty-midi/

On Sat, Jan 25, 2020 at 1:46 AM Ruoho Ruotsi notifications@github.com wrote:

Thanks CJ, I appreciate the tips! Once I'm going on the dataset, and have some initial generated basslines. I'll share it w/ you, perhaps there's a vector of collaboration here.

I haven't fine-tuned generative Transformers. But I have built/iterated/refined lots of other audio/dsp/speech models mostly on audio-signals though.

Do you have any recommendations on MIDI python libs/frameworks that'd be useful in whipping the MIDI data into shape? I fear I may have to manually pull the "riddim"/bassline out of the songs that are multitrack, others are just riddims + drum. Also debating whether to keep the "skank" with the bassline 🙇 🤔

— You are receiving this because you commented. Reply to this email directly, view it on GitHub https://github.com/RuohoRecords/Riddim-Archive/issues/3?email_source=notifications&email_token=AASXCZN6L6L2KSZ2EKVWZV3Q7PN2TA5CNFSM4IEVUD62YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOEJ4WW7A#issuecomment-578382716, or unsubscribe https://github.com/notifications/unsubscribe-auth/AASXCZO4LLQJEGKNVC4VXK3Q7PN2TANCNFSM4IEVUD6Q .