pytorch / pytorch

Tensors and Dynamic neural networks in Python with strong GPU acceleration
https://pytorch.org
Other
81.57k stars 21.89k forks source link

GPU Accelerated #120218

Closed TobaTV closed 6 months ago

TobaTV commented 6 months ago
### Tasks
TobaTV commented 6 months ago

The Fast Fourier Transform (FFT) calculates the Discrete Fourier Transform in O(n log n) time. It is foundational to a wide variety of numerical algorithms and signal processing techniques since it makes working in signals’ “frequency domains” as tractable as working in their spatial or temporal domains.

As part of PyTorch’s goal to support hardware-accelerated deep learning and scientific computing, we have invested in improving our FFT support, and with PyTorch 1.8, we are releasing the torch.fft module. This module implements the same functions as NumPy’s np.fft module, but with support for accelerators, like GPUs, and autograd.

GETTING STARTED Getting started with the new torch.fft module is easy whether you are familiar with NumPy’s np.fft module or not. While complete documentation for each function in the module can be found here, a breakdown of what it offers is:

fft, which computes a complex FFT over a single dimension, and ifft, its inverse the more general fftn and ifftn, which support multiple dimensions The “real” FFT functions, rfft, irfft, rfftn, irfftn, designed to work with signals that are real-valued in their time domains The “Hermitian” FFT functions, hfft and ihfft, designed to work with signals that are real-valued in their frequency domains Helper functions, like fftfreq, rfftfreq, fftshift, ifftshift, that make it easier to manipulate signals We think these functions provide a straightforward interface for FFT functionality, as vetted by the NumPy community, although we are always interested in feedback and suggestions!

To better illustrate how easy it is to move from NumPy’s np.fft module to PyTorch’s torch.fft module, let’s look at a NumPy implementation of a simple low-pass filter that removes high-frequency variance from a 2-dimensional image, a form of noise reduction or blurring:

import numpy as np import numpy.fft as fft

def lowpass_np(input, limit): pass1 = np.abs(fft.rfftfreq(input.shape[-1])) < limit pass2 = np.abs(fft.fftfreq(input.shape[-2])) < limit kernel = np.outer(pass2, pass1)

fft_input = fft.rfft2(input)
return fft.irfft2(fft_input * kernel, s=input.shape[-2:])

Now let’s see the same filter implemented in PyTorch:

import torch import torch.fft as fft

def lowpass_torch(input, limit): pass1 = torch.abs(fft.rfftfreq(input.shape[-1])) < limit pass2 = torch.abs(fft.fftfreq(input.shape[-2])) < limit kernel = torch.outer(pass2, pass1)

fft_input = fft.rfft2(input)
return fft.irfft2(fft_input * kernel, s=input.shape[-2:])

Not only do current uses of NumPy’s np.fft module translate directly to torch.fft, the torch.fft operations also support tensors on accelerators, like GPUs and autograd. This makes it possible to (among other things) develop new neural network modules using the FFT.

PERFORMANCE The torch.fft module is not only easy to use — it is also fast! PyTorch natively supports Intel’s MKL-FFT library on Intel CPUs, and NVIDIA’s cuFFT library on CUDA devices, and we have carefully optimized how we use those libraries to maximize performance. While your own results will depend on your CPU and CUDA hardware, computing Fast Fourier Transforms on CUDA devices can be many times faster than computing it on the CPU, especially for larger signals.

In the future, we may add support for additional math libraries to support more hardware. See below for where you can request additional hardware support.

Chillee commented 6 months ago

Closing for irrelevance