Published December 4, 2017 | Version v1
Conference paper

Some highlights on Source-to-Source Adjoint AD

Description

Algorithmic Differentiation (AD) provides the analytic derivatives of functions given as programs. Adjoint AD, which computes gradients, is similar to Back Propagation for Machine Learning. AD researchers study strategies to overcome the difficulties of adjoint AD, to get closer to its theoretical efficiency. To promote fruitful exchanges between Back Propagation and adjoint AD, we present three of these strategies and give our view of their interest and current status.

Abstract

International audience

Additional details

Created:
March 25, 2023
Modified:
November 29, 2023