Fast tensor operations using a convenient Einstein index notation.
Documentation | Digital Object Identifier | Downloads |
---|---|---|
Build Status | PkgEval | Coverage | Quality assurance |
---|---|---|---|
[][ci-url] | ![Codecov][codecov-img] |
[ci-url]: https://github.com/Jutho/TensorOperations.jl/actions?query=workflow%3ACI
[codecov-img]: https://codecov.io/gh/Jutho/TensorOperations.jl/branch/master/graph/badge.svg
Support for cuTENSOR v2 and with that more recent versions of CUDA.jl.
Improved support for automatic differentiation using reverse-mode rules with ChainRulesCore.jl.
Improved and extended support for backends and allocation strategies, with in particular support for allocating temporary objects using Bumper.jl.
Breaking changes to part of the interface to make it more sustainable for future improvements and extensions.
The @tensor
macro now accepts keyword arguments to facilitate a variety of options that help with debugging, contraction cost and backend selection.
Experimental support for automatic differentiation has been added by adding reverse-mode chainrules.
The interface for custom types has been changed and thoroughly documented, making it easier to know what to implement. This has as a consequence that more general element types of tensors are now also possible.
There is a new interface to work with backends, to allow for dynamic switching between different implementations of the primitive tensor operations or between different strategies for allocating new tensor objects.
The support for CuArray
objects is moved to a package extension, to avoid unnecessary CUDA dependencies for Julia versions >= 1.9
The cache for temporaries has been removed due to its inconsistent and intricate interplay with multithreading. However, the new feature of specifying custom allocation strategies can be used to experiment with novel cache-like behaviour in the future.
WARNING: TensorOperations 4.0 contains several breaking changes and cannot generally be expected to be compatible with previous versions.
TensorOperations.jl is mostly used through the @tensor
macro which allows one
to express a given operation in terms of
index notation format,
a.k.a. Einstein notation
(using Einstein's summation convention).
using TensorOperations
α = randn()
A = randn(5, 5, 5, 5, 5, 5)
B = randn(5, 5, 5)
C = randn(5, 5, 5)
D = zeros(5, 5, 5)
@tensor begin
D[a, b, c] = A[a, e, f, c, f, g] * B[g, b, e] + α * C[c, a, b]
E[a, b, c] := A[a, e, f, c, f, g] * B[g, b, e] + α * C[c, a, b]
end
In the second to last line, the result of the operation will be stored in the
preallocated array D
, whereas the last line uses a different assignment
operator :=
in order to define and allocate a new array E
of the correct
size. The contents of D
and E
will be equal.
For more detailed information, please see the documentation.
See CITATION.bib
for the relevant reference(s).