Section: Research Program
Algorithmic Differentiation
Participants : Laurent Hascoët, Valérie Pascual, Ala Taftaf.
-
Glossary
- algorithmic differentiation
(AD, aka Automatic Differentiation) Transformation of a program, that returns a new program that computes derivatives of the initial program, i.e. some combination of the partial derivatives of the program's outputs with respect to its inputs.
- adjoint
Mathematical manipulation of the Partial Derivative Equations that define a problem, obtaining new differential equations that define the gradient of the original problem's solution.
Algorithmic Differentiation (AD) differentiates
programs. The input of AD is
a source program
The resulting
where each
which can be mechanically written as a sequence of instructions
AD can be generalized to higher level derivatives, Taylor series, etc.
In practice, many applications only need cheaper projections of
-
Sensitivities, defined for a given direction
in the input space as:This expression is easily computed from right to left, interleaved with the original program instructions. This is the tangent mode of AD.
-
Adjoints, defined after transposition (
), for a given weighting of the outputs as:This expression is most efficiently computed from right to left, because matrix
vector products are cheaper than matrix matrix products. This defines the adjoint mode of AD, most effective for optimization, data assimilation [28] , adjoint problems [23] , or inverse problems.
Adjoint-mode AD turns out to make a very efficient program, at least
theoretically [25] . The computation time
required for the gradient is only a small multiple of the run-time of
However, the
Another research issue is to make the AD model cope with the constant evolution of modern language constructs. From the old days of Fortran77, novelties include pointers and dynamic allocation, modularity, structured data types, objects, vectorial notation and parallel communication. We keep developing our models and tools to handle these new constructs.