Open dajuno opened 2 years ago
This may be due to XLA. Could you try to disable XLA at the beginning of your code by:
dde.config.disable_xla_jit()
Yes, it works with XLA disabled. (And takes about 3x the time compared to tf1. still faster than pytorch, though)
DeepXDE uses different strategies of XLA for TensorFlow 1.x and TensorFlow 2.x. PyTorch is not supported by XLA in DeepXDE.
Hello, first, thanks for your excellent work.
Running the Lotka-Volterra Demo using DDE_BACKEND=tensorflow.compat.v1 works fine. Same program fails, when setting DDE_BACKEND=tensorflow. The computation seems to stall on the first iteration, GPU memory grows slowly, resulting in an out-of-memory error after a few minutes. During that process, there is CPU load but no GPU load (according to nvtop, only GPU memory usage). The computation runs without errors when commenting the lines
(although of course results are not meaningful.)
Versions used: deepxde 1.5.0 (from pypi) tensorflow 2.9 CUDA 11.2, cuDNN 8.4
Thank you!
======
Output with
tensorflow.compat.v1
:Output with
tensorflow
: