Closed skjerns closed 3 years ago
Thanks for the kind words! We are working on newer models for our narcolepsy classifier, but you are right, these results indicate a failure with the classification somewhere in the process.
Here are some things that may be contributing to the poor results:
Thanks for the quick response.
1) Unfortunately we do not have O1/O2 in all cases, often only O1 or O2 or Oz, or even only Pz 2) This is unfortunately not clear for me either: Our data is from different sources, so the labels are a bit mixed. What does the network expect? I would think referenced channels should be the input, right? 3) Ah, that is indeed the case, we have quite a bit of Wake before lights off. However, my intuition was, that the Hypnodensity is dealing with this quite well, as the accuracy score is above 75-80% (as compared to the F1 of 0.58).
Btw: If you are refactoring the model, it might be very useful to implement it in such a way that the models can be loaded once and be reused. I was able to adapt the NarcoApp
such that it loads the NC classification models and keeps them in memory for the succeeding classification, however, this was not possible for the Hypnodensity model (as variable names are the same and can't be that easily re-assigned in tensorflow).
You are welcome.
Thanks for the tip regarding keeping the models in memory. We are working to make the software more configurable for those who want to examine more than one study at a time, and will keep your thought here in mind.
3) I'll create new EDFs with the beginning cropped to lights off and run the analysis again - I'll let you know what comes out :) (will take a while, last time it took ~4-5 days for all calculations, however mainly the feature calculation takes a long time, and on the test machine I didn't have enough disk space to save the results unfortunately)
@skjerns Just to add a comment re. O1/O2: You do not explicitly need to have both, just one of them. The same applies for the central, right @informaton?
That is correct. Either O1 or O2 on its own is fine for the occipital lead, and either C3 or C4 for the central channel (like you said). Only one central and one occipital channel is fed into the model; the best quality occipital and central channels are selected whenever there are two to choose from. I don't know that we have fed Oz in though or what the sensitivity to the results are when using that or something like Pz in place of O1 or O2.
On Thu, Sep 3, 2020 at 7:41 AM Alexander Neergaard Olesen < notifications@github.com> wrote:
@skjerns https://github.com/skjerns Just to add a comment re. O1/O2: You do not explicitly need to have both, just one of them. The same applies for the central, right @informaton https://github.com/informaton?
— You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/Stanford-STAGES/stanford-stages/issues/22#issuecomment-686539666, or unsubscribe https://github.com/notifications/unsubscribe-auth/ABC3F7PVVWUV6DZF7WI7JBLSD6TKJANCNFSM4QMX6I5A .
Excellent work, and thanks so much that you are providing the model weights. Most reseachers don't supply the weights, which makes it sometimes impossible to recreate the results. So thank's again for doing so!
We're currently evaluating the model with a dataset we are working on of 100 NC1 patients and 130 controls. However, we're not able to reproduce the results reported in the paper.
Inter-rater reliability of hypnograms is okay-ish. However, I know myself how hard automatic sleep scoring is (mostly due to the broken scoring standard), so I think the results are overall good enough, and probably the best that are readily available on the open-source market.
However, the NC1 detection mostly fails for us
We're using the following input configuration
Do you have any tips on how to improve the performance? Should we add more channels (even if not entirely fitting, as Pz for O1?)?