NVlabs / stylegan3

Official PyTorch implementation of StyleGAN3
Other
6.28k stars 1.1k forks source link

filtered_lrelu called with parameters that have no optimized CUDA kernel, using generic fallback #622

Open ulucsahin opened 8 months ago

ulucsahin commented 8 months ago

Describe the bug A clear and concise description of what the bug is.

Output of _plugin.filtered_lrelu(x, fu, fd, b, si, up, down, px0, px1, py0, py1, sx, sy, gain, slope, clamp, flip_filter, write_signs) returns the return code -1, therefore it uses generic fallback. I want to use optimized kernels and not generic fallback. What might be the issue here for getting -1 return code?

I believe code enters

    if (!test_spec.exec)
    {
        // No kernel found - return empty tensors and indicate missing kernel with return code of -1.
        return std::make_tuple(torch::Tensor(), torch::Tensor(), -1);
    }

But I am not sure why this happens.

Desktop

ulucsahin commented 8 months ago

I disabled the code block in c++, and getting an issue as expected: internal error - CUDA kernel not found