# Integral of inverse functions

In mathematics, integrals of inverse functions can be computed by means of a formula that expresses the antiderivatives of the inverse $f^{-1}$ of a continuous and invertible function $f$, in terms of $f^{-1}$ and an antiderivative of $f$. This formula was published in 1905 by Charles-Ange Laisant.[1]

## Statement of the theorem

Let $I_1$ and $I_2$ be two intervals of $\mathbb{R}$. Assume that $f: I_1\to I_2$ is a continuous and invertible function, and let $f^{-1}$ denote its inverse $I_2\to I_1$. Then $f$ and $f^{-1}$ have antiderivatives, and if $F$ is an antiderivative of $f$, the possible antiderivatives of $f^{-1}$ are:

$\int f^{-1}(y)\,dy= y f^{-1}(y)-F\circ f^{-1}(y)+C,$

where $C$ is an arbitrary real number.

File:FunktionUmkehrIntegral2.svg
Proof without words of the theorem

In his 1905 article, Laisant gave three proofs. First, under the additional hypothesis that $f^{-1}$ is differentiable, one may differentiate the above formula, which completes the proof immediately. His second proof was geometric. If $f(a)=c$ and $f(b)=d$, the theorem can be written:

$\int_c^df^{-1}(y)\,dy+\int_a^bf(x)\,dx=bd-ac.$

The figure on the right is a proof without words of this formula. Laisant does not discuss the hypotheses necessary to make this proof rigorous, but it can be made explicit with the help of the Darboux integral[2] (or Fubini's theorem[3] if a demonstration based on the Lebesgue integral is desired). Laisant's third proof uses the additional hypothesis that $f$ is differentiable. Beginning with $f^{-1}(f(x)) = x$, one multiplies by $f'(x)$ and integrates both sides. The right-hand side is calculated using integration by parts to be $xf(x) - \textstyle\int f(x)\,dx$, and the formula follows.

Nevertheless, it can be shown that this theorem holds even if $f$ or $f^{-1}$ is not differentiable:[2][3] it suffices, for example, to use the Stieltjes integral in the previous argument. On the other hand, even though general monotonic functions are differentiable almost everywhere, the proof of the general formula does not follow, unless $f^{-1}$ is absolutely continuous.[3]

It is also possible to check that for every $y$ in $I_2$, the derivative of the function $y \to y f^{-1}(y) -F(f^{-1}(y))$ is equal to $f^{-1}(y)$.[4] In other words:

$\forall x\in I_1\quad\lim_{h\to 0}\frac{(x+h)f(x+h)-xf(x)-\left(F(x+h)-F(x)\right)}{f(x+h)-f(x)}=x.$

To this end, it suffices to apply the mean value theorem to $F$ between $x$ and $x+h$, taking into account that $f$ is monotonic.

## Examples

1. Assume that $f(x)=\exp(x)$, hence $f^{-1}(y)=\ln(y)$. The formula above gives immediately
$\int \ln(y) \, dy = y\ln(y)-y + C.$
2. Similarly, with $f(x)=\cos(x)$ and $f^{-1}(y)=\arccos(y)$,
$\quad\quad\int \arccos(y) \, dy = y\arccos(y) - \sin(\arccos(y))+C.$
3. With $\quad\quad f(x) = \tan(x)$ and $f^{-1}(y) = \arctan(y)$,
$\quad\quad\int \arctan(y) \, dy = y\arctan(y) + \ln|\cos(\arctan(y))| + C.$

## History

Apparently, this theorem of integration was discovered for the first time in 1905 by Charles-Ange Laisant,[1] who "could hardly believe that this theorem is new", and hoped its use would henceforth spread out among students and teachers. This result was published independently in 1912 by an Italian engineer, Alberto Caprilli, in an opuscule intitled "Nuove formole d'integrazione".[5] It was rediscovered in 1955 by Parker,[6] and by a number of mathematicians following him.[7] Nevertheless, they all assume that f or f-1 is differentiable. The general version of the theorem, free from this additional assumption, was proposed by Michael Spivak in 1965, as an exercise in the Calculus,[8] and a fairly complete proof following the same lines was published by Eric Key in 1994.[2] This proof relies on the very definition of the Darboux integral, and consists in showing that the upper Darboux sums of the function f are in 1-1 correspondence with the lower Darboux sums of f-1. In 2013, Michael Bensimhoun, estimating that the general theorem was still insufficiently known, gave two other proofs:[3] The second proof, based on the Stieltjes integral and on its formulae of integration by parts and of homeomorphic change of variables, is the most suitable to establish more complex formulae.[9]

## Generalization to holomorphic functions

The above theorem generalizes in the obvious way to holomorphic functions: Let $U$ and $V$ be two open and simply connected sets of $\mathbb{C}$, and assume that $f: U\to V$ is a biholomorphism. Then $f$ and $f^{-1}$ have antiderivatives, and if $F$ is an antiderivative of $f$, the general antiderivative of $f^{-1}$ is

$G(z)= z f^{-1}(z)-F\circ f^{-1}(z)+C.$

Because all holomorphic functions are differentiable, the proof is immediate by complex differentiation.

## References

1. Laisant, C.-A. (1905). "Intégration des fonctions inverses". Nouvelles annales de mathématiques, journal des candidats aux écoles polytechnique et normale. 5 (4): 253–257.
2. Key, E. (Mar 1994). "Disks, Shells, and Integrals of Inverse Functions". The College Mathematics Journal. 25 (2): 136–138. doi:10.2307/2687137.
3. Bensimhoun, Michael (2013). "On the antiderivative of inverse functions". arXiv:.
4. This very simple proof of the general theorem, the only one that does not make use of integrals, was communicated by the French mathematician and Wikipedian Anne Bauval in the corresponding pages in French. It seems to have escaped the persons who published proofs of this result.