API
Types and constants
EnzymeCore.ABIEnzymeCore.ActiveEnzymeCore.AnnotationEnzymeCore.BatchDuplicatedEnzymeCore.BatchDuplicatedNoNeedEnzymeCore.ConstEnzymeCore.DuplicatedEnzymeCore.DuplicatedNoNeedEnzymeCore.FFIABIEnzymeCore.ForwardModeEnzymeCore.InlineABIEnzymeCore.ModeEnzymeCore.ReverseModeEnzymeCore.ReverseModeSplitEnzymeCore.EnzymeRules.AugmentedReturnEnzymeCore.EnzymeRules.ConfigEnzymeTestUtils.ExprAndMsg
Functions and macros
Enzyme.gradientEnzyme.gradientEnzyme.gradientEnzyme.gradient!Enzyme.jacobianEnzyme.jacobianEnzymeCore.autodiffEnzymeCore.autodiffEnzymeCore.autodiffEnzymeCore.autodiffEnzymeCore.autodiff_deferredEnzymeCore.autodiff_deferredEnzymeCore.autodiff_deferredEnzymeCore.autodiff_deferred_thunkEnzymeCore.autodiff_thunkEnzymeCore.autodiff_thunkEnzymeCore.make_zeroEnzymeCore.make_zeroEnzymeCore.EnzymeRules.augmented_primalEnzymeCore.EnzymeRules.forwardEnzymeCore.EnzymeRules.inactiveEnzymeCore.EnzymeRules.inactive_noinlEnzymeCore.EnzymeRules.inactive_typeEnzymeCore.EnzymeRules.reverseEnzymeTestUtils.@test_msgEnzymeTestUtils.are_activities_compatibleEnzymeTestUtils.test_forwardEnzymeTestUtils.test_reverseEnzyme.API.runtimeActivityEnzyme.API.runtimeActivity!
Documentation
Enzyme.gradient! — Methodgradient!(::ReverseMode, dx, f, x)Compute the gradient of an array-input function f using reverse mode, storing the derivative result in an existing array dx.
Example:
f(x) = x[1]*x[2]
dx = [0.0, 0.0]
gradient!(Reverse, dx, f, [2.0, 3.0])
# output
2-element Vector{Float64}:
3.0
2.0Enzyme.gradient — Methodgradient(::ForwardMode, f, x; shadow=onehot(x))Compute the gradient of an array-input function f using forward mode. The optional keyword argument shadow is a vector of one-hot vectors of type x which are used to forward-propagate into the return. For performance reasons, this should be computed once, outside the call to gradient, rather than within this call.
Example:
f(x) = x[1]*x[2]
grad = gradient(Forward, f, [2.0, 3.0])
# output
(3.0, 2.0)Enzyme.gradient — Methodgradient(::ReverseMode, f, x)Compute the gradient of an array-input function f using reverse mode. This will allocate and return new array with the gradient result.
Example:
f(x) = x[1]*x[2]
grad = gradient(Reverse, f, [2.0, 3.0])
# output
2-element Vector{Float64}:
3.0
2.0Enzyme.gradient — Methodgradient(::ForwardMode, f, x, ::Val{chunk}; shadow=onehot(x))Compute the gradient of an array-input function f using vector forward mode. Like gradient, except it uses a chunk size of chunk to compute chunk derivatives in a single call.
Example:
f(x) = x[1]*x[2]
grad = gradient(Forward, f, [2.0, 3.0], Val(2))
# output
(3.0, 2.0)Enzyme.jacobian — Methodjacobian(::ForwardMode, f, x; shadow=onehot(x))
jacobian(::ForwardMode, f, x, ::Val{chunk}; shadow=onehot(x))Compute the jacobian of an array-input function f using (potentially vector) forward mode. This is a simple rename of the gradient function, and all relevant arguments apply here.
Example:
f(x) = [x[1]*x[2], x[2]]
grad = jacobian(Forward, f, [2.0, 3.0])
# output
2×2 Matrix{Float64}:
3.0 2.0
0.0 1.0Enzyme.jacobian — Methodjacobian(::ReverseMode, f, x, ::Val{num_outs}, ::Val{chunk})Compute the jacobian of an array-input function f using (potentially vector) reverse mode. The chunk argument denotes the chunk size to use and num_outs denotes the number of outputs f will return in an array.
Example:
f(x) = [x[1]*x[2], x[2]]
grad = jacobian(Reverse, f, [2.0, 3.0], Val(2))
# output
2×2 Matrix{Float64}:
3.0 2.0
0.0 1.0EnzymeCore.autodiff — Methodautodiff(::ForwardMode, f, Activity, args...)Auto-differentiate function f at arguments args using forward mode.
args may be numbers, arrays, structs of numbers, structs of arrays and so on. Enzyme will only differentiate in respect to arguments that are wrapped in a Duplicated or similar argument. Non-annotated arguments will automatically be treated as Const. Unlike reverse mode in autodiff, Active arguments are not allowed here, since all derivative results of immutable objects will be returned and should instead use Duplicated or variants like DuplicatedNoNeed.
Activity is the Activity of the return value, it may be:
Constif the return is not to be differentiated with respect toDuplicated, if the return is being differentiated with respect to and both the original value and the derivative return are desiredDuplicatedNoNeed, if the return is being differentiated with respect to and only the derivative return is desired.BatchDuplicated, likeDuplicated, but computing multiple derivatives at once. All batch sizes must be the same for all arguments.BatchDuplicatedNoNeed, likeDuplicatedNoNeed, but computing multiple derivatives at one. All batch sizes must be the same for all arguments.
Example returning both original return and derivative:
f(x) = x*x
res, ∂f_∂x = autodiff(Forward, f, Duplicated, Duplicated(3.14, 1.0))
# output
(9.8596, 6.28)Example returning just the derivative:
f(x) = x*x
∂f_∂x = autodiff(Forward, f, DuplicatedNoNeed, Duplicated(3.14, 1.0))
# output
(6.28,)EnzymeCore.autodiff — Methodautodiff(mode::Mode, f, args...)Like autodiff but will try to guess the activity of the return value.
EnzymeCore.autodiff — Methodautodiff(mode::Mode, f, ::Type{A}, args...)Like autodiff but will try to extend f to an annotation, if needed.
EnzymeCore.autodiff — Methodautodiff(::ReverseMode, f, Activity, args...)Auto-differentiate function f at arguments args using reverse mode.
Limitations:
fmay only return aReal(of a built-in/primitive type) ornothing, not an array, struct,BigFloat, etc. To handle vector-valued return types, use a mutatingf!that returnsnothingand stores it's return value in one of the arguments, which must be wrapped in aDuplicated.
args may be numbers, arrays, structs of numbers, structs of arrays and so on. Enzyme will only differentiate in respect to arguments that are wrapped in an Active (for arguments whose derivative result must be returned rather than mutated in place, such as primitive types and structs thereof) or Duplicated (for mutable arguments like arrays, Refs and structs thereof). Non-annotated arguments will automatically be treated as Const.
Activity is the Activity of the return value, it may be Const or Active.
Example:
a = 4.2
b = [2.2, 3.3]; ∂f_∂b = zero(b)
c = 55; d = 9
f(a, b, c, d) = a * √(b[1]^2 + b[2]^2) + c^2 * d^2
∂f_∂a, _, _, ∂f_∂d = autodiff(Reverse, f, Active, Active(a), Duplicated(b, ∂f_∂b), c, Active(d))[1]
# output
(3.966106403010388, nothing, nothing, 54450.0)here, autodiff returns a tuple $(\partial f/\partial a, \partial f/\partial d)$, while $\partial f/\partial b$ will be added to ∂f_∂b (but not returned). c will be treated as Const(c).
One can also request the original returned value of the computation.
Example:
Enzyme.autodiff(ReverseWithPrimal, x->x*x, Active(3.0))
# output
((6.0,), 9.0)Enzyme gradients with respect to integer values are zero. Active will automatically convert plain integers to floating point values, but cannot do so for integer values in tuples and structs.
EnzymeCore.autodiff_deferred — Methodautodiff_deferred(::ForwardMode, f, Activity, args...)Same as autodiff(::ForwardMode, ...) but uses deferred compilation to support usage in GPU code, as well as high-order differentiation.
EnzymeCore.autodiff_deferred — Methodautodiff_deferred(mode::Mode, f, ::Type{A}, args...)Like autodiff_deferred but will try to extend f to an annotation, if needed.
EnzymeCore.autodiff_deferred — Methodautodiff_deferred(::ReverseMode, f, Activity, args...)Same as autodiff but uses deferred compilation to support usage in GPU code, as well as high-order differentiation.
EnzymeCore.autodiff_deferred_thunk — Methodautodiff_deferred_thunk(::ReverseModeSplit, ftype, Activity, argtypes...)Provide the split forward and reverse pass functions for annotated function type ftype when called with args of type argtypes when using reverse mode.
Activity is the Activity of the return value, it may be Const, Active, or Duplicated (or its variants DuplicatedNoNeed, BatchDuplicated, and BatchDuplicatedNoNeed).
The forward function will return a tape, the primal (or nothing if not requested), and the shadow (or nothing if not a Duplicated variant), and tapes the corresponding type arguements provided.
The reverse function will return the derivative of Active arguments, updating the Duplicated arguments in place. The same arguments to the forward pass should be provided, followed by the adjoint of the return (if the return is active), and finally the tape from the forward pass.
Example:
A = [2.2]; ∂A = zero(A)
v = 3.3
function f(A, v)
res = A[1] * v
A[1] = 0
res
end
forward, reverse = autodiff_deferred_thunk(ReverseSplitWithPrimal, Const{typeof(f)}, Active, Duplicated{typeof(A)}, Active{typeof(v)})
tape, result, shadow_result = forward(Const(f), Duplicated(A, ∂A), Active(v))
_, ∂v = reverse(Const(f), Duplicated(A, ∂A), Active(v), 1.0, tape)[1]
result, ∂v, ∂A
# output
(7.26, 2.2, [3.3])EnzymeCore.autodiff_thunk — Methodautodiff_thunk(::ForwardMode, ftype, Activity, argtypes...)Provide the thunk forward mode function for annotated function type ftype when called with args of type argtypes.
Activity is the Activity of the return value, it may be Const or Duplicated (or its variants DuplicatedNoNeed, BatchDuplicated, andBatchDuplicatedNoNeed).
The forward function will return the primal (if requested) and the shadow (or nothing if not a Duplicated variant).
Example returning both original return and derivative:
a = 4.2
b = [2.2, 3.3]; ∂f_∂b = zero(b)
c = 55; d = 9
f(x) = x*x
forward = autodiff_thunk(Forward, Const{typeof(f)}, Duplicated, Duplicated{Float64})
res, ∂f_∂x = forward(Const(f), Duplicated(3.14, 1.0))
# output
(9.8596, 6.28)Example returning just the derivative:
a = 4.2
b = [2.2, 3.3]; ∂f_∂b = zero(b)
c = 55; d = 9
f(x) = x*x
forward = autodiff_thunk(Forward, Const{typeof(f)}, DuplicatedNoNeed, Duplicated{Float64})
∂f_∂x = forward(Const(f), Duplicated(3.14, 1.0))
# output
(6.28,)EnzymeCore.autodiff_thunk — Methodautodiff_thunk(::ReverseModeSplit, ftype, Activity, argtypes...)Provide the split forward and reverse pass functions for annotated function type ftype when called with args of type argtypes when using reverse mode.
Activity is the Activity of the return value, it may be Const, Active, or Duplicated (or its variants DuplicatedNoNeed, BatchDuplicated, and BatchDuplicatedNoNeed).
The forward function will return a tape, the primal (or nothing if not requested), and the shadow (or nothing if not a Duplicated variant), and tapes the corresponding type arguements provided.
The reverse function will return the derivative of Active arguments, updating the Duplicated arguments in place. The same arguments to the forward pass should be provided, followed by the adjoint of the return (if the return is active), and finally the tape from the forward pass.
Example:
A = [2.2]; ∂A = zero(A)
v = 3.3
function f(A, v)
res = A[1] * v
A[1] = 0
res
end
forward, reverse = autodiff_thunk(ReverseSplitWithPrimal, Const{typeof(f)}, Active, Duplicated{typeof(A)}, Active{typeof(v)})
tape, result, shadow_result = forward(Const(f), Duplicated(A, ∂A), Active(v))
_, ∂v = reverse(Const(f), Duplicated(A, ∂A), Active(v), 1.0, tape)[1]
result, ∂v, ∂A
# output
(7.26, 2.2, [3.3])EnzymeCore.ABI — Typeabstract type ABIAbstract type for what ABI will be used.
EnzymeCore.Active — TypeActive(x)Mark a function argument x of autodiff as active, Enzyme will auto-differentiate in respect Active arguments.
Enzyme gradients with respect to integer values are zero. Active will automatically convert plain integers to floating point values, but cannot do so for integer values in tuples and structs.
EnzymeCore.Annotation — Typeabstract type Annotation{T}Abstract type for autodiff function argument wrappers like Const, Active and Duplicated.
EnzymeCore.BatchDuplicated — TypeBatchDuplicated(x, ∂f_∂xs)Like Duplicated, except contains several shadows to compute derivatives for all at once. Argument ∂f_∂xs should be a tuple of the several values of type x.
EnzymeCore.BatchDuplicatedNoNeed — TypeBatchDuplicatedNoNeed(x, ∂f_∂xs)Like DuplicatedNoNeed, except contains several shadows to compute derivatives for all at once. Argument ∂f_∂xs should be a tuple of the several values of type x.
EnzymeCore.Const — TypeConst(x)Mark a function argument x of autodiff as constant, Enzyme will not auto-differentiate in respect Const arguments.
EnzymeCore.Duplicated — TypeDuplicated(x, ∂f_∂x)Mark a function argument x of autodiff as duplicated, Enzyme will auto-differentiate in respect to such arguments, with dx acting as an accumulator for gradients (so $\partial f / \partial x$ will be added to) ∂f_∂x.
EnzymeCore.DuplicatedNoNeed — TypeDuplicatedNoNeed(x, ∂f_∂x)Like Duplicated, except also specifies that Enzyme may avoid computing the original result and only compute the derivative values.
EnzymeCore.FFIABI — Typestruct FFIABI <: ABIForeign function call ABI. JIT the differentiated function, then inttoptr call the address.
EnzymeCore.ForwardMode — Typestruct Forward <: ModeForward mode differentiation
EnzymeCore.InlineABI — Typestruct InlineABI <: ABIInlining function call ABI.
EnzymeCore.Mode — Typeabstract type ModeAbstract type for what differentiation mode will be used.
EnzymeCore.ReverseMode — Typestruct ReverseMode{ReturnPrimal,ABI} <: Mode{ABI}Reverse mode differentiation.
ReturnPrimal: Should Enzyme return the primal return value from the augmented-forward.
EnzymeCore.ReverseModeSplit — Typestruct ReverseModeSplit{ReturnPrimal,ReturnShadow,Width,ModifiedBetween,ABI} <: Mode{ABI}Reverse mode differentiation.
ReturnPrimal: Should Enzyme return the primal return value from the augmented-forward.ReturnShadow: Should Enzyme return the shadow return value from the augmented-forward.Width: Batch Size (0 if to be automatically derived)ModifiedBetween: Tuple of each argument's modified between state (true if to be automatically derived).
EnzymeCore.make_zero — Functionmake_zero(::Type{T}, seen::IdDict, prev::T, ::Val{copy_if_inactive}=Val(false))::T
Recursively make a zero'd copy of the value `prev` of type `T`. The argument `copy_if_inactive` specifies
what to do if the type `T` is guaranteed to be inactive, use the primal (the default) or still copy the value.EnzymeCore.make_zero — Methodmake_zero(prev::T)
Helper function to recursively make zero.EnzymeCore.EnzymeRules.AugmentedReturn — TypeAugmentedReturn(primal, shadow, tape)Augment the primal return value of a function with its shadow, as well as any additional information needed to correctly compute the reverse pass, stored in tape.
Unless specified by the config that a variable is not overwritten, rules must assume any arrays/data structures/etc are overwritten between the forward and the reverse pass. Any floats or variables passed by value are always preserved as is (as are the arrays themselves, just not necessarily the values in the array).
See also augmented_primal.
EnzymeCore.EnzymeRules.Config — TypeConfig{NeedsPrimal, NeedsShadow, Width, Overwritten}
ConfigWidth{Width} = Config{<:Any,<:Any, Width}Configuration type to dispatch on in custom reverse rules (see augmented_primal and reverse).
NeedsPrimalandNeedsShadow: boolean values specifying whether the primal and shadow (resp.) should be returned.Width: an integer that specifies the number of adjoints/shadows simultaneously being propagated.Overwritten: a tuple of booleans of whether each argument (including the function itself) is modified between the forward and reverse pass (true if potentially modified between).
Getters for the four type parameters are provided by needs_primal, needs_shadow, width, and overwritten.
EnzymeCore.EnzymeRules.augmented_primal — Functionaugmented_primal(::Config, func::Annotation{typeof(f)}, RT::Type{<:Annotation}, args::Annotation...)Must return an AugmentedReturn type.
- The primal must be the same type of the original return if
needs_primal(config), otherwise nothing. - The shadow must be nothing if needs_shadow(config) is false. If width is 1, the shadow should be the same type of the original return. If the width is greater than 1, the shadow should be NTuple{original return, width}.
- The tape can be any type (including Nothing) and is preserved for the reverse call.
EnzymeCore.EnzymeRules.forward — Functionforward(func::Annotation{typeof(f)}, RT::Type{<:Annotation}, args::Annotation...)Calculate the forward derivative. The first argument func is the callable for which the rule applies to. Either wrapped in a Const), or a Duplicated if it is a closure. The second argument is the return type annotation, and all other arguments are the annotated function arguments.
EnzymeCore.EnzymeRules.inactive — Functioninactive(func::typeof(f), args...)Mark a particular function as always being inactive in both its return result and the function call itself.
EnzymeCore.EnzymeRules.inactive_noinl — Functioninactive_noinl(func::typeof(f), args...)Mark a particular function as always being inactive in both its return result and the function call itself, but do not prevent inlining of the function.
EnzymeCore.EnzymeRules.inactive_type — Methodinactive_type(::Type{Ty})Mark a particular type Ty as always being inactive.
EnzymeCore.EnzymeRules.reverse — Functionreverse(::Config, func::Annotation{typeof(f)}, dret::Active, tape, args::Annotation...)
reverse(::Config, func::Annotation{typeof(f)}, ::Type{<:Annotation), tape, args::Annotation...)Takes gradient of derivative, activity annotation, and tape. If there is an active return dret is passed as Active{T} with the derivative of the active return val. Otherwise dret is passed as Type{Duplicated{T}}, etc.
EnzymeTestUtils.ExprAndMsg — TypeA cunning hack to carry extra message along with the original expression in a test
EnzymeTestUtils.@test_msg — Macro@test_msg msg condion kws...This is per Test.@test condion kws... except that if it fails it also prints the msg. If msg=="" then this is just like @test, nothing is printed
Examles
julia> @test_msg "It is required that the total is under 10" sum(1:1000) < 10;
Test Failed at REPL[1]:1
Expression: sum(1:1000) < 10
Problem: It is required that the total is under 10
Evaluated: 500500 < 10
ERROR: There was an error during testing
julia> @test_msg "It is required that the total is under 10" error("not working at all");
Error During Test at REPL[2]:1
Test threw exception
Expression: error("not working at all")
Problem: It is required that the total is under 10
"not working at all"
Stacktrace:
julia> a = "";
julia> @test_msg a sum(1:1000) < 10;
Test Failed at REPL[153]:1
Expression: sum(1:1000) < 10
Evaluated: 500500 < 10
ERROR: There was an error during testingEnzymeTestUtils.are_activities_compatible — Methodare_activities_compatible(Tret, activities...) -> BoolReturn true if return activity type Tret and activity types activities are compatible.
EnzymeTestUtils.test_forward — Methodtest_forward(f, Activity, args...; kwargs...)Test Enzyme.autodiff of f in Forward-mode against finite differences.
f has all constraints of the same argument passed to Enzyme.autodiff, with several additional constraints:
- If it mutates one of its arguments, it must return that argument.
Arguments
Activity: the activity of the return value offargs: Each entry is either an argument tof, an activity type accepted byautodiff, or a tuple of the form(arg, Activity), whereActivityis the activity type ofarg. If the activity type specified requires a tangent, a random tangent will be automatically generated.
Keywords
fdm=FiniteDifferences.central_fdm(5, 1): The finite differences method to use.fkwargs: Keyword arguments to pass tof.rtol: Relative tolerance forisapprox.atol: Absolute tolerance forisapprox.testset_name: Name to use for a testset in which all tests are evaluated.
Examples
Here we test a rule for a function of scalars. Because we don't provide an activity annotation for y, it is assumed to be Const.
using Enzyme, EnzymeTestUtils
x, y = randn(2)
for Tret in (Const, Duplicated, DuplicatedNoNeed), Tx in (Const, Duplicated)
test_forward(*, Tret, (x, Tx), y)
endHere we test a rule for a function of an array in batch forward-mode:
x = randn(3)
y = randn()
for Tret in (Const, BatchDuplicated, BatchDuplicatedNoNeed),
Tx in (Const, BatchDuplicated),
Ty in (Const, BatchDuplicated)
test_forward(*, Tret, (x, Tx), (y, Ty))
endEnzymeTestUtils.test_reverse — Methodtest_reverse(f, Activity, args...; kwargs...)Test Enzyme.autodiff_thunk of f in ReverseSplitWithPrimal-mode against finite differences.
f has all constraints of the same argument passed to Enzyme.autodiff_thunk, with several additional constraints:
- If it mutates one of its arguments, it must not also return that argument.
- If the return value is a struct, then all floating point numbers contained in the struct or its fields must be in arrays.
Arguments
Activity: the activity of the return value off.args: Each entry is either an argument tof, an activity type accepted byautodiff, or a tuple of the form(arg, Activity), whereActivityis the activity type ofarg. If the activity type specified requires a shadow, one will be automatically generated.
Keywords
fdm=FiniteDifferences.central_fdm(5, 1): The finite differences method to use.fkwargs: Keyword arguments to pass tof.rtol: Relative tolerance forisapprox.atol: Absolute tolerance forisapprox.testset_name: Name to use for a testset in which all tests are evaluated.
Examples
Here we test a rule for a function of scalars. Because we don't provide an activity annotation for y, it is assumed to be Const.
using Enzyme, EnzymeTestUtils
x = randn()
y = randn()
for Tret in (Const, Active), Tx in (Const, Active)
test_reverse(*, Tret, (x, Tx), y)
endHere we test a rule for a function of an array in batch reverse-mode:
x = randn(3)
for Tret in (Const, Active), Tx in (Const, BatchDuplicated)
test_reverse(prod, Tret, (x, Tx))
endEnzyme.API.runtimeActivity! — MethodruntimeActivity!(val::Bool)Enzyme runs an activity analysis which deduces which values, instructions, etc are necessary to be differentiated and therefore involved in the differentiation procedure. This runs at compile time. However, there may be implementation flaws in this analysis that means that Enzyme cannot deduce that an inactive (const) value is actually const. Alternatively, there may be some data which is conditionally active, depending on which runtime branch is taken. In these cases Enzyme conservatively presumes the value is active.
However, in certain cases, an insufficiently aggressive activity analysis may result in derivative errors – for example by mistakenly using the primal (const) argument and mistaking it for the duplicated shadow. As a result this may result in incorrect results, or accidental updates to the primal.
This flag enables runntime activity which tells all load/stores to check at runtime whether the value they are updating is indeed active (in addition to the compile-time activity analysis). This will remedy these such errors, but at a performance penalty of performing such checks.
It is on the Enzyme roadmap to add a PotentiallyDuplicated style activity, in addition to the current Const and Duplicated styles that will disable the need for this, which does not require the check when a value is guaranteed active, but still supports runtime-based activity information.
This function takes an argument to set the runtime activity value, true means it is on, and false means off. By default it is off.
Enzyme.API.runtimeActivity — MethodruntimeActivity()Gets the current value of the runtime activity. See runtimeActivity! for more information.