Open ChrisRackauckas opened 8 years ago
Yes I think it would be the right thing to do to allow b
as a matrix, so that the Krylov methods we have will generalize automatically to block Krylov methods. The main thing I don't know how to do is what termination criteria to write down. It's tempting to simply replace all the norms with matrix norms, but I'm not sure if that's the right thing to do.
It's tempting to simply replace all the norms with matrix norms, but I'm not sure if that's the right thing to do.
Note that if this is done, we may need to require packages like GenericSVD.jl in order to support number types like BigFloat since the SVD is used for the 2-norm and no generic fallback is given in Base.
Hi,
Has anyone made progress on this issue? In my case, I need to solve a 2d PDE with matrix-free method. I represent the solution as a matrix. The Frobenium norm would do for the stopping criterion.
Just to be sure, I think what you mean is support for custom vector types, right? Block methods are about solving a system AX=B for multiple right-hand sides, in the sense that you solve AX[:, i] = B[:, i] for all i's independently. This is not the same as representing the unknowns in a shape similar to your geometry (then you still have a single rhs).
In principle the iterative method is just unaware about the geometry of your domain, so you could just do the transformation from vector -> your internal geometric representation of the unknowns before applying your matrix-free A, and then transforming it back again to vector. And in fact, this should be a no-op, since the underlying data structure should be the same. It's just reinterpreting a vector.
You are right, my 2d-array is like a vector. It is not meant for solving block systems AX[:, i] = B[:, i]
.
I though about using reshape
but fear about performance loss. This is also not very convenient if the data is hold on the GPU and no good reshape
method is available.
For example, if I take A = x-> fft(x,2)
, how can I use solve Ax=y
using your package?
It's a general problem. See also https://github.com/JuliaLang/julia/issues/25565. In Julia, we have generally associated vector with 1d arrays and linear maps with 2d arrays but, as you mention, it doesn't have to be like that. It just makes things simpler. It looks like people would like a broader vector definition to happen for dot/inner
and norm
but next up is axpy!
, and should we generally allow A*x
as long as size(A, 2) == length(x)
even though x
is 2d or 3d.
I'm on Julia 0.7 right now, so I can't really test this approximately working Julia 0.6 code:
import Base: A_mul_B!
struct MyLinearOp
n::Int
end
function A_mul_B!(y, A::MyLinearOp, x)
fft!(reshape(copy!(y, x), A.n, A.n), 2) # copy x -> y and reshape to a matrix, then do fft; reshape shouldn't copy iirc
y # return the original y, not the reshaped reference to it
end
function example(n = 128)
A = MyLinearOp(n)
x_square = rand(Complex128, n, n)
y_square = similar(x)
A_mul_B!(y_square, A, x_square) # should work
x_linear = rand(Complex128, n * n)
y_linear = similar(x_linear)
A_mul_B!(y_linear, A, x_linear) # should work
# then just call
b = ... # some vector
x = gmres(A, b)
# and transform x back to your matrix-like representation.
end
But it might complain you have to implement one or two things, such as eltype, size etc. A better way to work with matrix-free methods is to use https://github.com/Jutho/LinearMaps.jl
Thank you for your help!!
With LinearMaps
, I dont think it is possible without the reshape because the constructor requires the input size N
to be an Int
and not a tuple for example.
LinearMap{T=Float64}(f, [fc,], M::Int, N::Int = M; kwargs...)
Basically, my question is "Can we solve Ax = y" without reshaping and y being a 2d array?
Sorry, I think I got it now...
Basically, my question is "Can we solve Ax = y" without reshaping and y being a 2d array?
reshaping is just a view so it's free.
Indeed. There is a related example here at LinearMaps.jl
, which works just fine on Julia 0.7. I hadn't done any performance testing though. When using LinearMaps
with the FunctionMap
type, you should get full compliance with IterativeSolvers
automatically, no need to redefine LinearAlgebra
methods etc.
Late to the party, but here is some code for testing:
using LinearAlgebra, LinearMaps, BenchmarkTools
# with the following tweak, one may define multiplication for arrays
function LinearAlgebra.mul!(y::AbstractMatrix, A::LinearMaps.FunctionMap, x::AbstractMatrix)
(length(x) == A.N && length(y) == A.M) || throw(DimensionMismatch())
LinearMaps.ismutating(A) ? A.f(y, x) : copyto!(y, A.f(x))
return y
end
using DSP, FFTW
n = 1024;
A = LinearMap((y, x) -> (FFTW.fft!(reshape(copyto!(y, x), n, n), 2); y),
(y, x) -> (FFTW.ifft!(reshape(copyto!(y, x), n, n), 2); y),
n*n)
B = LinearMap((y, x) -> FFTW.fft!(copyto!(y, x), 2),
(y, x) -> FFTW.ifft!(copyto!(y, x), 2),
n*n)
x = rand(ComplexF64, n*n);
y = similar(x);
@benchmark mul!($y, $A, $x)
x = rand(ComplexF64, n, n);
y = similar(x);
@benchmark mul!($y, $B, $x)
This yields
julia> @benchmark mul!($y, $A, $x)
BenchmarkTools.Trial:
memory estimate: 2.83 KiB
allocs estimate: 45
--------------
minimum time: 17.217 ms (0.00% GC)
median time: 17.857 ms (0.00% GC)
mean time: 18.095 ms (0.00% GC)
maximum time: 24.932 ms (0.00% GC)
--------------
samples: 277
evals/sample: 1
for the first benchmark, and
julia> @benchmark mul!($y, $B, $x)
BenchmarkTools.Trial:
memory estimate: 2.83 KiB
allocs estimate: 45
--------------
minimum time: 17.178 ms (0.00% GC)
median time: 18.130 ms (0.00% GC)
mean time: 18.334 ms (0.00% GC)
maximum time: 25.287 ms (0.00% GC)
--------------
samples: 273
evals/sample: 1
for the second. Doesn't seem to be so bad.
For when mul!(Y, A, X)
is meant to apply A
columnwise to X::AbstractMatrix
, there is a corresponding function mul!
in LinearMaps.jl
.
thank you!! Sorry for the late reply but this does not seem to work
sol = IterativeSolvers.gmres(B, x)
I was wondering if you guys would think it's necessary to support arbitrary sized inputs. For example, instead of just allowing matrix vector, allow matrix matrix in the normal interpretation. This can come up in solving PDEs (see DifferentialEquations.jl). Currently the issue with the CG and GMRES functions are due to their use of the
dot
function, and I could change it up if people think this is the right way to go.