OMEinsum.DynamicEinCodeType
DynamicEinCode{LT}
DynamicEinCode(ixs, iy)

Wrapper to eincode-specification that creates a callable object to evaluate the eincode ixs -> iy where ixs are the index-labels of the input-tensors and iy are the index-labels of the output.

example

julia> a, b = rand(2,2), rand(2,2);

julia> OMEinsum.DynamicEinCode((('i','j'),('j','k')),('i','k'))(a, b) ≈ a * b
true
source
OMEinsum.DynamicNestedEinsumType
DynamicNestedEinsum{LT} <: NestedEinsum{LT}
DynamicNestedEinsum(args, eins)
DynamicNestedEinsum{LT}(tensorindex::Int)

Einsum with contraction order, where the type parameter LT is the label type. It has two constructors. One takes a tensorindex as input, which represents the leaf node in a contraction tree. The other takes an iterable of type DynamicNestedEinsum, args, as the siblings, and eins to specify the contraction operation.

source
OMEinsum.EinArrayType
EinArray{T, N, TT, LX, LY, ICT, OCT} <: AbstractArray{T, N}

A struct to hold the intermediate result of an einsum where all index-labels of both input and output are expanded to a rank-N-array whose values are lazily calculated. Indices are arranged as inner indices (or reduced dimensions) first and then outer indices.

Type parameters are

* `T`: element type,
* `N`: array dimension,
* `TT`: type of "tuple of input arrays",
* `LX`: type of "tuple of input indexers",
* `LX`: type of output indexer,
* `ICT`: typeof inner CartesianIndices,
* `OCT`: typeof outer CartesianIndices,
source
OMEinsum.EinCodeType
EinCode <: AbstractEinsum
EinCode(ixs, iy)

Abstract type for sum-product contraction code. The constructor returns a DynamicEinCode instance.

source
OMEinsum.EinIndexerType
EinIndexer{locs,N}

A structure for indexing EinArrays. locs is the index positions (among all indices). In the constructor, size is the size of target tensor,

source
OMEinsum.EinIndexerMethod
EinIndexer{locs}(size::Tuple)

Constructor for EinIndexer for an object of size size where locs are the locations of relevant indices in a larger tuple.

source
OMEinsum.IndexGroupType
IndexGroup

Leaf in a contractiontree, contains the indices and the number of the tensor it describes, e.g. in "ij,jk -> ik", indices "ik" belong to tensor 1, so would be described by IndexGroup(['i','k'], 1).

source
OMEinsum.NestedEinsumConstructorType
NestedEinsumConstructor

describes a (potentially) nested einsum. Important fields:

  • args, vector of all inputs, either IndexGroup objects corresponding to tensors or NestedEinsumConstructor
  • iy, indices of output
source
OMEinsum.SlicedEinsumType
SlicedEinsum{LT, Ein} <: AbstractEinsum

A tensor network with slicing. LT is the label type and Ein is the tensor network.

Fields

  • slicing::Vector{LT}: A vector of labels to slice.
  • eins::Ein: The tensor network.
source
OMEinsum.StaticEinCodeType
StaticEinCode{LT, ixs, iy}

The static version of DynamicEinCode that matches the contraction rule at compile time. It is the default return type of @ein_str macro. LT is the label type.

source
OMEinsum.StaticNestedEinsumType
StaticNestedEinsum{LT,args,eins} <: NestedEinsum{LT}
StaticNestedEinsum(args, eins)
StaticNestedEinsum{LT}(tensorindex::Int)

Einsum with contraction order, where the type parameter LT is the label type, args is a tuple of StaticNestedEinsum, eins is a StaticEinCode and leaf node is defined by setting eins to an integer. It has two constructors. One takes a tensorindex as input, which represents the leaf node in a contraction tree. The other takes an iterable of type DynamicNestedEinsum, args, as the siblings, and eins to specify the contraction operation.

source
Base.getindexMethod
getindex(A::EinArray, inds...)

return the lazily calculated entry of A at index inds.

source
OMEinsum.allow_loopsMethod
allow_loops(flag::Bool)

Setting this to false will cause OMEinsum to log an error if it falls back to loop_einsum evaluation, instead of calling specialised kernels. The default is true.

source
OMEinsum.alluniqueMethod
allunique(ix::Tuple)

return true if all elements of ix appear only once in ix.

example

julia> using OMEinsum: allunique

julia> allunique((1,2,3,4))
true

julia> allunique((1,2,3,1))
false
source
OMEinsum.asarrayMethod
asarray(x[, parent::AbstractArray]) -> AbstactArray

Return a 0-dimensional array with item x, otherwise, do nothing. If a parent is supplied, it will try to match the parent array type.

source
OMEinsum.back_propagateMethod
back_propagate(f, code, cache, ȳ, size_dict)

Back propagate the message through the cached tree cache and return a tree storing the intermediate messages. The message can be gradients et al.

Arguments

  • f: The back-propagation rule. The signature is f(eins, xs, y, size_dict, dy) -> dxs, where
    • eins: The contraction code at the current node.
    • xs: The input tensors at the current node.
    • y: The output tensor at the current node.
    • size_dict: The size dictionary, which maps the label to the size of the corresponding dimension.
    • dy: The message on the output tensor (y) to back-propagate through the current node.
    • dxs: The message on the input tensors (xs) as the result of back-propagation.
  • code: The contraction code, which can be a NestedEinsum or a SlicedEinsum.
  • cache: The cached intermediate results, which can be generated by cached_einsum.
  • : The message to back-propagate.
  • size_dict: The size dictionary, which maps the label to the size of the corresponding dimension.

Returns

  • CacheTree: The tree storing the intermediate messages.
source
OMEinsum.cached_einsumMethod
cached_einsum(code, xs, size_dict)

Compute the einsum contraction and cache the intermediate contraction results.

Arguments

  • code: The contraction code, which can be a NestedEinsum or a SlicedEinsum.
  • xs: The input tensors.
  • size_dict: The size dictionary, which maps the label to the size of the corresponding dimension.

Returns

  • CacheTree: The cached tree storing the intermediate results.
source
OMEinsum.cost_and_gradientFunction
cost_and_gradient(code, xs, ȳ)

Compute the cost and the gradients w.r.t the input tensors xs.

Arguments

  • code: The contraction code, which can be a NestedEinsum or a SlicedEinsum.
  • xs: The input tensors.
  • : The message to back-propagate. Default is 1.

Returns

  • cost: The cost of the contraction.
  • grads: The gradients w.r.t the input tensors.
source
OMEinsum.einarrayMethod
einarray(::Val{ixs}, Val{iy}, xs, size_dict) -> EinArray

Constructor of EinArray from an EinCode, a tuple of tensors xs and a size_dict that assigns each index-label a size. The returned EinArray holds an intermediate result of the einsum specified by the EinCode with indices corresponding to all unique labels in the einsum. Reduction over the (lazily calculated) dimensions that correspond to labels not present in the output lead to the result of the einsum.

example

julia> using OMEinsum: get_size_dict

julia> a, b = rand(2,2), rand(2,2);

julia> sd = get_size_dict((('i','j'),('j','k')), (a, b));

julia> ea = OMEinsum.einarray(Val((('i','j'),('j','k'))),Val(('i','k')), (a,b), sd);

julia> dropdims(sum(ea, dims=1), dims=1) ≈ a * b
true
source
OMEinsum.einsumFunction
einsum(code::EinCode, xs, size_dict)

Return the tensor that results from contracting the tensors xs according to the contraction code code.

Arguments

  • code: The einsum notation, which can be an instance of EinCode, NestedEinsum, or SlicedEinsum.
  • xs - the input tensors
  • size_dict - a dictionary that maps index-labels to their sizes

Examples

julia> a, b = rand(2,2), rand(2,2);

julia> einsum(EinCode((('i','j'),('j','k')),('i','k')), (a, b)) ≈ a * b
true

julia> einsum(EinCode((('i','j'),('j','k')),('k','i')), (a, b)) ≈ permutedims(a * b, (2,1))
true
source
OMEinsum.einsum!Function
einsum!(code::EinCode, xs, y, sx, sy, size_dict)

Inplace version of einsum. The result is stored in y.

Arguments

  • code: The einsum notation, which can be an instance of EinCode, NestedEinsum, or SlicedEinsum.
  • xs: The input tensors.
  • y: The output tensor.
  • sx: Scale x by sx.
  • sy: Scale y by sy.
  • size_dict: A dictionary that maps index-labels to their sizes.
source
OMEinsum.einsum_gradMethod
einsum_grad(ixs, xs, iy, size_dict, cdy, i)

return the gradient of the result of evaluating the EinCode w.r.t the ith tensor in xs. cdy is the result of applying the EinCode to the xs.

example

julia> using OMEinsum: einsum_grad, get_size_dict

julia> a, b = rand(2,2), rand(2,2);

julia> c = einsum(EinCode((('i','j'),('j','k')), ('i','k')), (a,b));

julia> sd = get_size_dict((('i','j'),('j','k')), (a,b));

julia> einsum_grad((('i','j'),('j','k')), (a,b), ('i','k'), sd, c, 1) ≈ c * transpose(b)
true
source
OMEinsum.filliys!Method
filliys!(neinsum::NestedEinsumConstructor)

goes through all NestedEinsumConstructor objects in the tree and saves the correct iy in them.

source
OMEinsum.get_size_dict!Method
get_size_dict!(ixs, xs, size_info)

return a dictionary that is used to get the size of an index-label in the einsum-specification with input-indices ixs and tensors xs after consistency within ixs and between ixs and xs has been verified.

source
OMEinsum.getixsvMethod
getixsv(code)

Get labels of input tensors for EinCode, NestedEinsum and some other einsum like objects. Returns a vector of vectors.

julia> getixsv(ein"(ij,jk),k->i")
3-element Vector{Vector{Char}}:
 ['i', 'j']
 ['j', 'k']
 ['k']
source
OMEinsum.getiyvMethod
getiy(code)

Get labels of the output tensor for EinCode, NestedEinsum and some other einsum like objects. Returns a vector.

julia> getiyv(ein"(ij,jk),k->i")
1-element Vector{Char}:
 'i': ASCII/Unicode U+0069 (category Ll: Letter, lowercase)
source
OMEinsum.indices_and_locsMethod
indices_and_locs(ixs,iy)

given the index-labels of input and output of an einsum, return (in the same order):

  • a tuple of the distinct index-labels of the output iy
  • a tuple of the distinct index-labels in ixs of the input not appearing in the output iy
  • a tuple of tuples of locations of an index-label in the ixs in a list of all index-labels
  • a tuple of locations of index-labels in iy in a list of all index-labels

where the list of all index-labels is simply the first and the second output catenated and the second output catenated.

source
OMEinsum.loop_einsum!Method
loop_einsum!(ixs, iy, xs, y, sx, sy, size_dict)

inplace-version of loop_einsum, saving the result in a preallocated tensor of correct size y.

source
OMEinsum.loop_einsumMethod
loop_einsum(::EinCode, xs, size_dict)

evaluates the eincode specified by EinCode and the tensors xs by looping over all possible indices and calculating the contributions ot the result. Scales exponentially in the number of distinct index-labels.

source
OMEinsum.map_prodMethod
map_prod(xs, ind, indexers)

calculate the value of an EinArray with EinIndexers indexers at location ind.

source
OMEinsum.match_ruleMethod
match_rule(ixs, iy)
match_rule(code::EinCode)

Returns the rule that matches, otherwise use DefaultRule - the slow loop_einsum backend.

source
OMEinsum.nopermuteMethod
nopermute(ix,iy)

check that all values in iy that are also in ix have the same relative order,

example

julia> using OMEinsum: nopermute

julia> nopermute((1,2,3),(1,2))
true

julia> nopermute((1,2,3),(2,1))
false

e.g. nopermute((1,2,3),(1,2)) is true while nopermute((1,2,3),(2,1)) is false

source
OMEinsum.parse_parensMethod
parse_parens(s::AbstractString, i, narg)

parse one level of parens starting at index i where narg counts which tensor the current group of indices, e.g. "ijk", belongs to. Recursively calls itself for each new opening paren that's opened.

source
OMEinsum.@ein!Macro
@ein! A[i,k] := B[i,j] * C[j,k]     # A = B * C
@ein! A[i,k] += B[i,j] * C[j,k]     # A += B * C

Macro interface similar to that of other packages.

Inplace version of @ein.

example

julia> a, b, c, d = rand(2,2), rand(2,2), rand(2,2), zeros(2,2);

julia> cc = copy(c);

julia> @ein! d[i,k] := a[i,j] * b[j,k];

julia> d ≈ a * b
true

julia> d ≈ ein"ij,jk -> ik"(a,b)
true

julia> @ein! c[i,k] += a[i,j] * b[j,k];

julia> c ≈ cc + a * b
true
source
OMEinsum.@einMacro
@ein A[i,k] := B[i,j] * C[j,k]     # A = B * C

Macro interface similar to that of other packages.

You may use numbers in place of letters for dummy indices, as in @tensor, and need not name the output array. Thus A = @ein [1,2] := B[1,ξ] * C[ξ,2] is equivalent to the above. This can also be written A = ein"ij,jk -> ik"(B,C) using the numpy-style string macro.

example

julia> a, b = rand(2,2), rand(2,2);

julia> @ein c[i,k] := a[i,j] * b[j,k];

julia> c ≈ a * b
true

julia> c ≈ ein"ij,jk -> ik"(a,b)
true
source
OMEinsum.@ein_strMacro
ein"ij,jk -> ik"(A,B)

String macro interface which understands numpy.einsum's notation. Translates strings into StaticEinCode-structs that can be called to evaluate an einsum. To control evaluation order, use parentheses - instead of an EinCode, a NestedEinsum is returned which evaluates the expression according to parens. The valid character ranges for index-labels are a-z and α-ω.

example

julia> a, b, c = rand(10,10), rand(10,10), rand(10,1);

julia> ein"ij,jk,kl -> il"(a,b,c) ≈ ein"(ij,jk),kl -> il"(a,b,c) ≈ a * b * c
true
source
OMEinsum.@optein_strMacro
optein"ij,jk,kl -> ik"(A, B, C)

String macro interface that similar to @ein_str, with optimized contraction order (dimensions are assumed to be uniform).

source
OMEinsumContractionOrders.AbstractEinsumType
AbstractEinsum

Abstract type for einsum notations.

Required Interfaces

  • getixsv: a vector of vectors, each vector represents the labels associated with a input tensor.
  • getiyv: a vector of labels associated with the output tensor.
  • uniquelabels: a vector of labels that are unique in the einsum notation.

Derived interfaces

  • labeltype: the data type to represent the labels in the einsum notation.
source
OMEinsumContractionOrders.BipartiteResultType
BipartiteResult{RT}
BipartiteResult(part1, part2, sc, valid)

Result of the bipartite optimization. part1 and part2 are the two parts of the bipartition, sc is the space complexity of the bipartition, valid is a boolean indicating whether the bipartition is valid.

source
OMEinsumContractionOrders.EinCodeType
EinCode{LT} <: AbstractEinsum
EinCode(ixs::Vector{Vector{LT}}, iy::Vector{LT})

Einsum code with input indices ixs and output index iy.

Examples

The einsum notation for matrix multiplication is:

julia> code = OMEinsumContractionOrders.EinCode([[1,2], [2, 3]], [1, 3])
1∘2, 2∘3 -> 1∘3

julia> OMEinsumContractionOrders.getixsv(code)
2-element Vector{Vector{Int64}}:
 [1, 2]
 [2, 3]

julia> OMEinsumContractionOrders.getiyv(code)
2-element Vector{Int64}:
 1
 3
source
OMEinsumContractionOrders.ExactTreewidthType
const ExactTreewidth = Treewidth{SafeRules{BT, MMW{3}(), MF}}
ExactTreewidth() = Treewidth()

ExactTreewidth is a specialization of Treewidth for the SafeRules preprocessing algorithm with the BT elimination algorithm. The BT algorithm is an exact solver for the treewidth problem that implemented in TreeWidthSolver.jl.

source
OMEinsumContractionOrders.GreedyMethodType
GreedyMethod{MT}
GreedyMethod(; α = 0.0, temperature = 0.0)

It may not be optimal, but it is fast.

Fields

  • α is the parameter for the loss function, for pairwise interaction, L = size(out) - α * (size(in1) + size(in2))
  • temperature is the parameter for sampling, if it is zero, the minimum loss is selected; for non-zero, the loss is selected by the Boltzmann distribution, given by p ~ exp(-loss/temperature).
source
OMEinsumContractionOrders.HyperNDType
HyperND(;
    dis = KaHyParND(),
    algs = (MF(), AMF(), MMD()),
    level = 6,
    width = 120,
    imbalances = 130:130,
    score = ScoreFunction(),
)

Nested-dissection based optimizer. Recursively partitions a tensor network, then calls a greedy algorithm on the leaves. The optimizer is run a number of times: once for each greedy algorithm in algs and each imbalance value in imbalances. The recursion depth is controlled by the parameters level and width.

The line graph is partitioned using the algorithm dis. OMEinsumContractionOrders currently supports two partitioning algorithms, both of which require importing an external library.

typepackage
METISNDMetis.jl
KaHyParNDKayHyPar.jl

The optimizer is implemented using the tree decomposition library CliqueTrees.jl.

Arguments

source
OMEinsumContractionOrders.KaHyParBipartiteType
KaHyParBipartite{RT,IT,GM}
KaHyParBipartite(; sc_target, imbalances=collect(0.0:0.005:0.8),
    max_group_size=40, greedy_config=GreedyMethod())

Optimize the einsum code contraction order using the KaHyPar + Greedy approach. This program first recursively cuts the tensors into several groups using KaHyPar, with maximum group size specifed by max_group_size and maximum space complexity specified by sc_target, Then finds the contraction order inside each group with the greedy search algorithm. Other arguments are

Fields

  • sc_target is the target space complexity, defined as log2(number of elements in the largest tensor),
  • imbalances is a KaHyPar parameter that controls the group sizes in hierarchical bipartition,
  • max_group_size is the maximum size that allowed to used greedy search,
  • sub_optimizer is the sub-optimizer used to find the contraction order when the group size is small enough.

References

source
OMEinsumContractionOrders.MergeGreedyType
MergeGreedy <: CodeSimplifier
MergeGreedy(; threshhold=-1e-12)

Contraction code simplifier (in order to reduce the time of calling optimizers) that merges tensors greedily if the space complexity of merged tensors is reduced (difference smaller than the threshhold).

source
OMEinsumContractionOrders.NestedEinsumType
NestedEinsum{LT} <: AbstractEinsum
NestedEinsum(args::Vector{NestedEinsum}, eins::EinCode)

The einsum notation with a contraction order specified as a tree data structure. It is automatically generated by the contraction code optimizer with the optimize_code function.

Fields

  • args: the children of the current node
  • tensorindex: the index of the input tensor, required only for leaf nodes. For non-leaf nodes, it is -1.
  • eins: the einsum notation for the operation at the current node.
source
OMEinsumContractionOrders.SABipartiteType
SABipartite{RT,BT}
SABipartite(; sc_target=25, ntrials=50, βs=0.1:0.2:15.0, niters=1000
    max_group_size=40, greedy_config=GreedyMethod(), initializer=:random)

Optimize the einsum code contraction order using the Simulated Annealing bipartition + Greedy approach. This program first recursively cuts the tensors into several groups using simulated annealing, with maximum group size specifed by max_group_size and maximum space complexity specified by sc_target, Then finds the contraction order inside each group with the greedy search algorithm. Other arguments are

Fields

  • sc_target is the target space complexity, defined as log2(number of elements in the largest tensor),
  • ntrials is the number of repetition (with different random seeds),
  • βs is a list of inverse temperature 1/T,
  • niters is the number of iteration in each temperature,
  • max_group_size is the maximum size that allowed to used greedy search,
  • sub_optimizer is the optimizer for the bipartited sub graphs, one can choose GreedyMethod() or TreeSA(),
  • initializer is the partition configuration initializer, one can choose :random or :greedy (slow but better).

References

source
OMEinsumContractionOrders.ScoreFunctionType
ScoreFunction

A function to compute the score of a contraction code:

score = tc_weight * 2^tc + rw_weight * 2^rw + sc_weight * max(0, 2^sc - 2^sc_target)

Fields

  • tc_weight: the weight of the time complexity, default is 1.0.
  • sc_weight: the weight of the space complexity (the size of the largest tensor), default is 1.0.
  • rw_weight: the weight of the read-write complexity, default is 0.0.
  • sc_target: the target space complexity, below which the sc_weight will be set to 0 automatically, default is 0.0.
source
OMEinsumContractionOrders.SlicedEinsumType
SlicedEinsum{LT,ET<:Union{EinCode{LT},NestedEinsum{LT}}} <: AbstractEinsum
SlicedEinsum(slicing::Vector{LT}, eins::ET)

The einsum notation with sliced indices. The sliced indices are the indices enumerated manually at the top level. By slicing the indices, the space complexity of the einsum notation can be reduced.

Fields

  • slicing: the sliced indices.
  • eins: the einsum notation of the current node, which is a NestedEinsum object.
source
OMEinsumContractionOrders.TreeSAType
TreeSA{IT} <: CodeOptimizer
TreeSA(; βs=collect(0.01:0.05:15), ntrials=10, niters=50, initializer=:greedy, score=ScoreFunction())

Optimize the einsum contraction pattern using the simulated annealing on tensor expression tree.

Fields

  • ntrials, βs and niters are annealing parameters, doing ntrials indepedent annealings, each has inverse tempteratures specified by βs, in each temperature, do niters updates of the tree.
  • initializer specifies how to determine the initial configuration, it can be :greedy, :random or :specified. If the initializer is :specified, the input code should be a NestedEinsum object.
  • score specifies the score function to evaluate the quality of the contraction tree, it is a function of time complexity, space complexity and read-write complexity.

References

Breaking changes:

  • nslices is removed, since the slicing part is now separated from the optimization part, see slice_code function and TreeSASlicer.
  • greedy_method is removed. If you want to have detailed control of the initializer, please pre-optimize the code with another method and then use :specified to initialize the tree.
source
OMEinsumContractionOrders.TreeSASlicerType
TreeSASlicer{IT, LT} <: CodeSlicer

A structure for configuring the Tree Simulated Annealing (TreeSA) slicing algorithm. The goal of slicing is to reach the target space complexity specified by score.sc_target.

Fields

  • ntrials, βs and niters are annealing parameters, doing ntrials indepedent annealings, each has inverse tempteratures specified by βs, in each temperature, do niters updates of the tree.
  • fixed_slices::Vector{LT}: A vector of fixed slices that should not be altered. Default is an empty vector.
  • optimization_ratio::Float64: A constant used for determining the number of iterations for slicing. Default is 2.0. i.e. if the current space complexity is 30, and the target space complexity is 20, then the number of iterations for slicing is (30 - 20) x optimization_ratio.
  • score::ScoreFunction: A function to evaluate the quality of the contraction tree. Default is ScoreFunction(sc_target=30.0).

References

source
OMEinsumContractionOrders.TreewidthType
struct Treewidth{EL <: EliminationAlgorithm, GM} <: CodeOptimizer
Treewidth(; alg::EL = SafeRules(BT(), MMW{3}(), MF()))

Tree width based solver. The solvers are implemented in CliqueTrees.jl and TreeWidthSolver.jl. They include:

AlgorithmDescriptionTime ComplexitySpace Complexity
BFSbreadth-first searchO(m + n)O(n)
MCSmaximum cardinality searchO(m + n)O(n)
LexBFSlexicographic breadth-first searchO(m + n)O(m + n)
RCMMDreverse Cuthill-Mckee (minimum degree)O(m + n)O(m + n)
RCMGLreverse Cuthill-Mckee (George-Liu)O(m + n)O(m + n)
MCSMmaximum cardinality search (minimal)O(mn)O(n)
LexMlexicographic breadth-first search (minimal)O(mn)O(n)
AMFapproximate minimum fillO(mn)O(m + n)
MFminimum fillO(mn²)-
MMDmultiple minimum degreeO(mn²)O(m + n)

Detailed descriptions is available in the CliqueTrees.jl.

Fields

  • alg::EL: The algorithm to use for the treewidth calculation. Available elimination algorithms are listed above.

Example

julia> optimizer = Treewidth();

julia> eincode = OMEinsumContractionOrders.EinCode([['a', 'b'], ['a', 'c', 'd'], ['b', 'c', 'e', 'f'], ['e'], ['d', 'f']], ['a'])
ab, acd, bcef, e, df -> a

julia> size_dict = Dict([c=>(1<<i) for (i,c) in enumerate(['a', 'b', 'c', 'd', 'e', 'f'])]...)
Dict{Char, Int64} with 6 entries:
  'f' => 64
  'a' => 2
  'c' => 8
  'd' => 16
  'e' => 32
  'b' => 4

julia> optcode = optimize_code(eincode, size_dict, optimizer)
ba, ab -> a
├─ bcf, fac -> ba
│  ├─ e, bcef -> bcf
│  │  ├─ e
│  │  └─ bcef
│  └─ df, acd -> fac
│     ├─ df
│     └─ acd
└─ ab
source
OMEinsumContractionOrders.contraction_complexityMethod
contraction_complexity(eincode, size_dict) -> ContractionComplexity

Returns the time, space and read-write complexity of the einsum contraction. The returned ContractionComplexity object contains 3 fields:

  • tc: time complexity defined as log2(number of element-wise multiplications).
  • sc: space complexity defined as log2(size of the maximum intermediate tensor).
  • rwc: read-write complexity defined as log2(the number of read-write operations).
source
OMEinsumContractionOrders.embed_simplifierMethod
embed_simplifier(code::NestedEinsum, simplifier::NetworkSimplifier)

Embed the simplifier into the contraction code. A typical workflow is: (i) generate a simplifier with simplify_code, (ii) then optimize the simplified code with optimize_code and (iii) post-process the optimized code with embed_simplifier to produce correct contraction order for the original code. This is automatically done in optimize_code given the simplifier argument is not nothing.

Arguments

  • code: the contraction code to embed the simplifier into.
  • simplifier: the simplifier to embed, which is a NetworkSimplifier object.

Returns

  • A new NestedEinsum object.
source
OMEinsumContractionOrders.flopMethod
flop(eincode, size_dict) -> Int

Returns the number of iterations, which is different with the true floating point operations (FLOP) by a factor of 2.

source
OMEinsumContractionOrders.getixsvFunction
getixsv(code::AbstractEinsum) -> Vector{Vector{LT}}

Returns the input indices of the einsum notation. Each vector represents the labels associated with a input tensor.

source
OMEinsumContractionOrders.label_elimination_orderMethod
label_elimination_order(code) -> Vector

Returns a vector of labels sorted by the order they are eliminated in the contraction tree. The contraction tree is specified by code, which e.g. can be a NestedEinsum instance.

source
OMEinsumContractionOrders.optimize_codeMethod
optimize_code(eincode, size_dict, optimizer = GreedyMethod(); slicer=nothing, simplifier=nothing, permute=true) -> optimized_eincode

Optimize the einsum contraction code and reduce the time/space complexity of tensor network contraction. Returns a NestedEinsum instance. Input arguments are

Arguments

  • eincode is an einsum contraction code instance, one of DynamicEinCode, StaticEinCode or NestedEinsum.
  • size is a dictionary of "edge label=>edge size" that contains the size information, one can use uniformsize(eincode, 2) to create a uniform size.
  • optimizer is a CodeOptimizer instance, should be one of GreedyMethod, Treewidth, KaHyParBipartite, SABipartite or TreeSA. Check their docstrings for details.

Keyword Arguments

  • slicer is for slicing the contraction code to reduce the space complexity, default is nothing. Currently only TreeSASlicer is supported.
  • simplifier is one of MergeVectors or MergeGreedy. Default is nothing.
  • permute is a boolean flag to indicate whether to optimize the permutation of the contraction order.

Examples

julia> using OMEinsum

julia> code = ein"ij, jk, kl, il->"
ij, jk, kl, il -> 

julia> optimize_code(code, uniformsize(code, 2), TreeSA());
source
OMEinsumContractionOrders.optimize_greedyMethod
optimize_greedy(eincode, size_dict; α, temperature)

Greedy optimizing the contraction order and return a NestedEinsum object. Check the docstring of tree_greedy for detailed explaination of other input arguments.

source
OMEinsumContractionOrders.optimize_saMethod
optimize_sa(code, size_dict; sc_target, max_group_size=40, βs=0.1:0.2:15.0, niters=1000, ntrials=50,
       sub_optimizer = GreedyMethod(), initializer=:random)

Optimize the einsum code contraction order using the Simulated Annealing bipartition + Greedy approach. size_dict is a dictionary that specifies leg dimensions. Check the docstring of SABipartite for detailed explaination of other input arguments.

References

source
OMEinsumContractionOrders.optimize_treeMethod
optimize_tree(code, size_dict; βs, ntrials, niters, initializer, score)

Optimize the einsum contraction pattern specified by code, and edge sizes specified by size_dict. Check the docstring of TreeSA for detailed explaination of other input arguments.

source
OMEinsumContractionOrders.optimize_treewidthMethod
optimize_treewidth(optimizer, eincode, size_dict)

Optimizing the contraction order via solve the exact tree width of the line graph corresponding to the eincode and return a NestedEinsum object. Check the docstring of treewidth_method for detailed explaination of other input arguments.

source
OMEinsumContractionOrders.simplify_codeMethod
simplify_code(code::Union{EinCode, NestedEinsum}, size_dict, method::CodeSimplifier)

Simplify the contraction code by preprocessing the code with a simplifier.

Arguments

  • code: the contraction code to simplify.
  • size_dict: the size dictionary of the contraction code.
  • method: the simplifier to use, which can be MergeVectors or MergeGreedy.

Returns

  • A tuple of (NetworkSimplifier, newcode), where newcode is a new EinCode object.
source
OMEinsumContractionOrders.slice_codeMethod
slice_code(code, size_dict, slicer) -> sliced_code

Slice the einsum contraction code to reduce the space complexity, returns a SlicedEinsum instance.

Arguments

  • code is a NestedEinsum instance.
  • size_dict is a dictionary of "edge label=>edge size" that contains the size information, one can use uniformsize(eincode, 2) to create a uniform size.
  • slicer is a CodeSlicer instance, currently only TreeSASlicer is supported.
source
OMEinsumContractionOrders.tree_greedyMethod
tree_greedy(incidence_list, log2_sizes; α = 0.0, temperature = 0.0)

Compute greedy order, and the time and space complexities, the rows of the incidence_list are vertices and columns are edges. log2_sizes are defined on edges. α is the parameter for the loss function, for pairwise interaction, L = size(out) - α * (size(in1) + size(in2)) temperature is the parameter for sampling, if it is zero, the minimum loss is selected; for non-zero, the loss is selected by the Boltzmann distribution, given by p ~ exp(-loss/temperature).

source
OMEinsumContractionOrders.viz_contractionMethod
viz_contraction(code::Union{NestedEinsum, SlicedEinsum}; locs=StressLayout(), framerate=10, filename=tempname() * ".mp4", show_progress=true)

Visualize the contraction process of a tensor network.

Arguments

  • code: The tensor network to visualize.

Keyword Arguments

  • locs: The coordinates or layout algorithm to use for positioning the nodes in the graph. Default is StressLayout().
  • framerate: The frame rate of the animation. Default is 10.
  • filename: The name of the output file, with .gif or .mp4 extension. Default is a temporary file with .mp4 extension.
  • show_progress: Whether to show progress information. Default is true.

Returns

  • the path of the generated file.
source
OMEinsumContractionOrders.viz_einsMethod
viz_eins(code::AbstractEinsum; locs=StressLayout(), filename = nothing, kwargs...)

Visualizes an AbstractEinsum object by creating a tensor network graph and rendering it using GraphViz.

Arguments

  • code::AbstractEinsum: The AbstractEinsum object to visualize.

Keyword Arguments

  • locs=StressLayout(): The coordinates or layout algorithm to use for positioning the nodes in the graph.
  • filename = nothing: The name of the file to save the visualization to. If nothing, the visualization will be displayed on the screen instead of saving to a file.
  • config = GraphDisplayConfig(): The configuration for displaying the graph. Please refer to the documentation of GraphDisplayConfig for more information.
  • kwargs...: Additional keyword arguments to be passed to the GraphViz constructor.
source