Closed amhummer closed 3 years ago
Hi @amhummer, it has been days since the last commit. Sorry that I probably need some time to get familiar with my code again. I will respond by tonight.
Hi @amhummer, thank you for opening the thread, but sorry that we may not be able to find the exact pre-trained model that 'result_human_test.txt' corresponds to. This work was collaborated by two institutions. We shared the common code base, but verified results independently to ensure the report being reliable. Since we have large population of samples for testing, a result was valid if different runs (with randomness) were statistically consistent.
model_tmp.npy
was put in the repo for tutorial purpose while lstm_0.589547.npy
was actually uploaded by mistake. I will soon add it to .gitignore
and put a note in README. Sorry for any caused confusion. By the way, the error you encountered was caused by the inconsistency between the latest code and the pre-trained model which was based on an earlier code version and didn't have an attribute named gapped
.
Our approach intended to offer as many flexibilities as possible for different needs. Researchers may prefer aligning their sequences by different schemes while some others want to apply algorithms on raw sequences directly. That's another reason why we didn't have it in mind to provide pre-trained models.
Hope this reply could be helpful. Be a bit off the topic, would you please tell me why you try reproducing the exact result?
Hi Alissa, I should have a copy of that pre-trained model that I used for the paper if you'd like to have it. Based on the comment above it sounds like there's a chance it may not be compatible with the current version of the code, but you can always try it and see. Let me know if you'd like me to send along that file.
Cheers, Andrew
Hi @xf3227 and @SenyorDrew,
Many thanks for your quick and helpful responses.
Andrew – It would be great if you could share that file with me, thank you very much! How would work best for you to send it (e.g., via GitHub or something else)?
@xf3227 – I am benchmarking humanisation methods on a new dataset and as one of the best performing and most-recently published tools, I wanted to ensure your LSTM model is represented. I just wanted to check that I am achieving the recorded level of performance on your test data and as reported in your paper before trying it on my dataset.
Many thanks, Alissa
Hi @amhummer Alissa, you are welcomed!
Greetings! @SenyorDrew Andrew, great to have you helping Alissa out! According to the error that Alissa reported, the previously saved model might be inconsistent with the latest code. There could potentially be more errors considering that the code structure was reworked for publishing on github. @amhummer when you have Andrew's model and encounter any code related issues, just let me know, I will guide you through to make it work.
Hi Alissa,
I've attached a tarred gzipped file. You should be able to extract it using:
tar -xzf peds_data.gz
You'll get a directory called "share" that contains:
Hope this helps!
Cheers, Andrew
Hi Andrew, This is very helpful – thank you so much for taking the time! Best, Alissa
Hi,
Thank you very much for providing the code for your model!
I am trying to reproduce your results for this LSTM method for calculating antibody humanness, as presented in Wollacott et al., 2019.
Could you confirm whether the model on your GitHub is identical to the one used in the Wollacott paper and, if so, how to evaluate test data identically to how you did in the paper? If it is not the same model, could you please advise me how to tune the model on GitHub to obtain the same results reported in the paper and/or provide me with access to the pre-trained model that achieved the results reported in the paper?
I have tried to run the model on the human_test.txt data provided using two approaches based on your documentation (full code included below). With one approach (
python ablstm.py eval … model_tmp.npy …
), I obtain similar but not identical test set results to what is reported in the result_human_test.txt file. With the other approach (from ablstm import ModelLSTM…
), I obtain much higher/poorer scores than expected or reported (~3, instead of ~0-1).Any guidance on how to reproduce the model and results from the paper would be greatly appreciated. Thank you very much in advance for your time and help!
Best wishes, Alissa Hummer
My code to run the model on the provided test data:
(1) In the command line:
python3 ablstm.py eval data/sample/human_test.txt saved_models/tmp/model_tmp.npy human_test-output.txt
(2) In a Python script:
Also, I am unable to run the saved_models/tmp/lstm_0.589547.npy model, as I encounter the following error: