There is this nice problem about sequences which I've encountered several times while solving problems in real analysis.

Two constant real numbers $a,b$ are given and then we have this sequence defined as:

$$a_0 = a$$

$$a_1 = b$$

$$a_{n+2} = \frac{a_{n+1} + a_n}{2},\ \ \ n \ge 0$$

Prove that the sequence converges and find the limit $L = \lim_{n \to \infty} a_n$

I won't post the solution here but... it turns out the limit is this number

$$L = \frac{1}{3} \cdot a + \frac{2}{3} \cdot b $$

Here is a nice illustration of this fact generated by a Python program.

In this case (depicted on the picture) the limit is $$\frac{1}{3} \cdot 10 + \frac{2}{3} \cdot 100 = \frac{1}{3} (10 + 200) = \frac{210}{3} = 70$$

## No comments:

## Post a Comment