Published December 4, 2017
| Version v1
Conference paper
Some highlights on Source-to-Source Adjoint AD
- Creators
- Hascoët, Laurent
Description
Algorithmic Differentiation (AD) provides the analytic derivatives of functions given as programs. Adjoint AD, which computes gradients, is similar to Back Propagation for Machine Learning. AD researchers study strategies to overcome the difficulties of adjoint AD, to get closer to its theoretical efficiency. To promote fruitful exchanges between Back Propagation and adjoint AD, we present three of these strategies and give our view of their interest and current status.
Abstract
International audience
Additional details
- URL
- https://hal.inria.fr/hal-01655085
- URN
- urn:oai:HAL:hal-01655085v1
- Origin repository
- UNICA