Closed tobigithub closed 5 years ago
Thank you for the detailed logs! Our environment uses CUDA 9.2. @tkschmidt and me are at a conference from today to mid June. We will look into that when we are back. Sorry for the inconvenience! :(
Thank you.
When thinking about this error ValueError: Tensor("out_target:0", shape=(?, ?), dtype=float32) must be from the same graph as Tensor("out/Reshape:0", shape=(?, ?), dtype=float32).
it occurred to me, that this may indicate that the model did not load properly.
Please try 'make jump MODEL=/path/to/the/model/. You should get an interactive bash with a
~/model/where you can find the model mounted. It could be that Docker is confused because there is no
/` at the model directories path.
Please try 'make jump MODEL=/path/to/the/model/
. You should get an interactive bash with a
~/model/where you can find the model mounted. It could be that Docker is confused because there is no
/` at the model directories path.
Tried that, it works, I can see all the files and HDF5, but loading the server with our without "/" gives the same error. What is the next step for "jump", writing a little CSV reader to predict the file inside docker?
coming across the same issue. I was using cuda9.0
@tony-jy-zhao Please understand that we cannot troubleshoot individual cuda versions and the errors they are causing.
@tobigithub you can use ipython to interactively step through the script that would usually called by make
. No need to write a parser. '/examples/' has example csv files that we tested and worked.
Specifically run:
from prosit import constants
from prosit import model as model_lib
model_dir = constants.MODEL_DIR
global model
global model_config
model, model_config = model_lib.load(model_dir, trained=True)
The server script is here: https://github.com/kusterlab/prosit/blob/master/prosit/server.py
As make jump is working this seems to be unrelated to docker versions. More likely a duplicate of #2
@gessulat thanks, the jump config works with CUDA 10, the server breaks. Will open new issue.
Hi, I had to upgrade/dowgrade docker and nvidia-docker because of missing images and version for ubuntu16 and currently cuda10. For the current version docker versions on https://download.docker.com/linux/ubuntu/dists/xenial/pool/edge/amd64/ there was no matching nvidia-docker-container so basically I pinned specific versions and installed 18.03 and NVIDIA Docker: 2.0.3
Docker and the nvidia-docker2 run fine.
and
The installation runs fine, i also can use nvidia-smi inside docker
I have downloaded the prosit1 model with config.yml, model.yml and weight_32_0.10211.hdf5 however when I run make server MODEL=/home/xxx/prosit/prosit1/ the server will start and greet me, but uploading a file with curl will do so, but then break. curl -F "peptides=@examples/peptidelist.csv" http://127.0.0.1:5000/predict/
Not sure how to debug this could be Keras of TF incompatibility with CUDA10? However I have Keras and TF running with different versions successfully outside docker. Tobias