napulen / phd_thesis

Automatic Roman numeral analysis in symbolic music representations.
1 stars 1 forks source link

[AugmentedNet] Learning about added convolutional blocks instead of concatenated #25

Closed napulen closed 2 years ago

napulen commented 2 years ago

As per the paper, I concatenate the Bass and Chroma convolutional blocks. I never tried adding them instead concatenating them.

The main difference is that adding results in a smaller network (less parameters), and the performance could be very similar.

I ran two experiments,

  1. one with an Add operation and the same Dense layer configuration
  2. one with an Add operation and a smaller Dense layer configuration

1 seemed either slightly worse or not different than the Concatenate version 2 seems definitely worse than the Concatenate version

napulen commented 2 years ago

The sizes of the network are:

Concatenate: 94,679 trainable parameters Add1: 89,431 trainable parameters Add2: 84,567 trainable parameters

One could make a case for saving those 5k parameters in the Add1 variation, as long as the performance is not detrimental. I think it is not worth it and it's best to keep the Concatenate.

napulen commented 2 years ago

I spoke too soon, Add2 seemed to have provided similar or slightly better results than Concatenate after the full 100 epochs.

Not in all tasks, but in several of them.

Here is the mlflow comparison. Left is Add2, right is Concatenate

image

napulen commented 2 years ago

Sticking to Concatenate for now. Also removed Dropout. The network is on its vanilla state now.