srvk / DiViMe

ACLEW Diarization Virtual Machine
Apache License 2.0
32 stars 9 forks source link

yunitator with 2 classes: target child vs. all other speakers #16

Closed alecristia closed 6 years ago

alecristia commented 6 years ago

that is, distinguish key child (one class) versus all others (another class containing mother, all female adults, male adults, other children, none)

alecristia commented 6 years ago

datasets that could be used for this:

All of the following need to be processed:

riebling commented 6 years ago

Hopefully this is trivial; you could imagine just filtering the RTTM produced by Yunitator with sed and awk, replacing class labels other than CHI from the set of, as I recall:

CHI
FEM
MAL
SIL
fmetze commented 6 years ago

I think the main idea is to train a new Yunitator on "ACLEW round 1-3" data in addition to Vandam and ASE (ACLEW Starter) - not sure if CHILDES and IDS/ADS samples are available.

alecristia commented 6 years ago

I second Eric's comment -- it might be good to get started with the data you already have, even if we'll increase the training set later on (which might make for a nice experiment anyway). Remember that the data folks are waiting for this tool to do their next round of sampling.

All of the data I mentioned are already available. All that remains is to format the annotations in the same format you have used before. Here are the links:

CHILDES Paidologos:

IDS/ADS samples:

riebling commented 6 years ago

Oh good. I also just had to request the latest media.talkbank.org data password, and can supply if needed.

riebling commented 6 years ago

This is maybe the second situation where we'd like to train up a new Yunitator variant. Maybe we can generalize and create a task for Yun: produce documentation and give examples of how we ("a novice") can train Yunitator on new data to produce new models & class labels?

alecristia commented 6 years ago

Closing this issue - this is a suboptimal solution, and we should focus our forces on better ones (eg 4- or 5-class labeling.)

riebling commented 6 years ago

I meant that if we can re-train a new Yunitator, it could be on 4- or 5-class labeled data, making it more optimal. Agree the number of classes currently produced by Yunitator is suboptimal.

alecristia commented 6 years ago

Oh I meant the 2-class solution that I had brought up in this issue was even more suboptimal (even less optimal?). It was just an idea that came up given the scarcity of data -- but I think we can do better, for instance with the 3 class currently implemented. RE training, I don't think that is a priority given our current user base (ACLEW), since none of us independently have enough data for retraining. In fact, it seems that even all together we don't have enough data for training 3 classes! So a retrainable module sounds more useful in theory than in our current world...