Closed lucas0 closed 5 years ago
This system is designed to adapt to new domains using only the sentences from the new domain, and does not require specificity labels from the new domain. The target['test']['path'] is used only for evaluation purposes, and does not effect the result, so you don't need to generate them. Those labels were annotated on MTurk.
Thank you very much. Are those calls the right way to do it for a new domain?
python train.py --gpu_id 0 --test_data my_data
python test.py --gpu_id 0 --test_data my_data
Also, how would the model work for longer sentences? Is it ok to put a paragraph with multiple sentences as a single input? Would the varying length of the paragraphs be an issue?
Yes, It's the right way. It is ok to use multiple sentences as a single input. I have only run this on the released datasets and some dialog datasets, so I am not sure how it will work if the inputs are very long.
In your instructions you say:
to change
s1['test']['path']
ands1['unlab']['path']
but there are two other variables (target['test']['path']
andtarget['test']['path']
) that hold the target values depending on the data you're using to train/test your model.What should we assign to those? From your instructions seems that we should use the target values for twitter/yelp/movies for our dataset.
How have you arrived on these values?
My data2.py and train.py/test.py modifications are like this:
And then the calls are:
This is working, but I'm not comfortable using the twitter
target['test']['path']
andtarget['test']['path']
. Is there any way I could generate those values for my data?Thanks a lot for the work and making the code available. This will help me a lot in my work.
Best, Lucas.