@austinmoehle Just as your explanation in the notebook, cudnn version gru is very fast than the tf implementation. Do you know if it's possible to use cudnn version gru when inference?
In your pb model, it seems that you use while loop rather than cudnn version gru.
@austinmoehle Just as your explanation in the notebook, cudnn version gru is very fast than the tf implementation. Do you know if it's possible to use cudnn version gru when inference?
In your pb model, it seems that you use while loop rather than cudnn version gru.