Closed n-zeng closed 2 years ago
Hi n-zeng! If I understand correctly, your machine learning model is severely under classifying the presence of your behavior: every frame is scored as behavior absent. The probabilities that the behavior is present are also very low for all frames.
One way this could happen would be if you fed a lot of annotations with behavior absent into your classifier, together with a very small number of behavior present annotations. How many annotated frames with the behavior-present vs behavior-absent did you use to create the model?
If the annotations are imbalanced towards behavior-absent, this would bias the model towards classifying most frames as behavior absent. One way around this is to balance your annotations, taking an equal or similar amount of behavior-absent and behavior-present frames when you create your classifier – check out the Random undersampling in the machine model settings for how to balance the data.
Describe the bug When analyzing the machine-results CSV's (C:\Users\name\Desktop\name\project_folder\csv\machine_results), I kept seeing all 0's in the column where it should be either a 0 or 1 based on whether the behavior is actually occurring (which I believe are derived from targets-inserted / Boris file ethograms). I'm also getting unexpectedly low results in the behavior probability columns.
To Reproduce
I was testing a previous and functional model on new data, so tell me if I did something incorrectly.
Steps to reproduce the behavior:
Expected behavior I was expecting to see the 0/1 column to match the data found in my ethograms, and for the probabilities to match that.
Additional context Unfortunately Github isn't letting me upload a screenshot right now, please let me know if anything needs to be clarified.