Open azhe825 opened 7 years ago
thanks for watching the ltierature
"Each round, besides all the labeled examples, randomly sample from the unlabeled examples and treat them as negative training examples." i like the idea.
It is simple, but effective.
important that you should stop soon walking circles in new land till you document the land you have visited. you need to get our 2 more papers, quick smart. high priority.
Sure. Please create a blank sharelatex project for me. I will fill in the rest.
Newest result in e-discovery: Scalability of Continuous Active Learning for Reliable High-Recall Text Classification mentioned one technique to tackle the problem.
Presumptive non-relevant examples. Autonomy and reliability of continuous active learning for technology-assisted review
May be useful for REUSE.
Testing.
What
Each round, besides all the labeled examples, randomly sample from the unlabeled examples and treat them as negative training examples.
Then train the model.
Why
E-discovery
why we need this technique:
why it works:
SLR
why we need this technique:
why it works:
Results
FASTREAD, use this tech or not:
Hall:
Wahono:
Abdellatif:
At least as good as not using it. (worst case result depends on pseudo random, not reliable)
Transfer learning result with this tech:
Hall as previous SLR,
on Wahono:
on Abdellatif:
Wahono as previous SLR,
on Hall:
on Abdellatif:
Abdellatif as previous SLR,
on Hall:
on Wahono:
Conclusions