Closed Baran-phys closed 3 months ago
When I use x-transformers (continuous models), my cached memory tend to surge during training. Also, the model uses 600-800% CPU. Is this due to different cloning or some other reasons?
When I use x-transformers (continuous models), my cached memory tend to surge during training. Also, the model uses 600-800% CPU. Is this due to different cloning or some other reasons?