Hi, @Meteorix ,
Could you please give a tensorflow example? For example, here is a pb model named "model.pb" and the input tensor is "input_x:0" and the output tensor is "logits:0", how can I use service-streamer to encapsulate it? The following is the code of the model prediction:
But I can never successfully realizztion the batch infer function. As you can see, the TensorFlow model will use GPU memory when it was loaded, so how can I assign the cuda devices? I've struggled in this problem for many days, could you please give me some help? Thanks a lot.
Hi, @Meteorix , Could you please give a tensorflow example? For example, here is a pb model named "model.pb" and the input tensor is "input_x:0" and the output tensor is "logits:0", how can I use service-streamer to encapsulate it? The following is the code of the model prediction:
asr_path='./model/model.pb' def load_model(model_path):
class AsrModel(object):
class ManagedAsrModel(ManagedModel):
But I can never successfully realizztion the batch infer function. As you can see, the TensorFlow model will use GPU memory when it was loaded, so how can I assign the cuda devices? I've struggled in this problem for many days, could you please give me some help? Thanks a lot.