Grasping the concept of Taylor Remainder Theorem

The name of the pictureThe name of the pictureThe name of the pictureClash Royale CLAN TAG#URR8PPP











up vote
0
down vote

favorite












I am trying to grasp the idea of Taylor Remainder Theorem. I want to know the way I understand it is right or wrong.



Like in the linear approximation



it is $$f(x) approx f(c) + f'(c)(x-c)$$
or $$f(x) = f(c) + f'(c)(x-c) + f''(zeta)(x-c)$$



which is in the picture below
enter image description here



The error is exactly $$f''(zeta)(x-c)$$



Then I thought of a case with a lower order of derivative which is



$$f(x) = f(c) + f'(zeta)(x-c)$$



enter image description here



The error is exactly $$f'(zeta)(x-c)$$



Then I noticed as there are more terms the order of derivatives for error gets higher which leads to the equation $$fracf^n+1(zeta)(n+1)!(x-a)^n+1$$



Is this the right way of understanding the theorem?







share|cite|improve this question


























    up vote
    0
    down vote

    favorite












    I am trying to grasp the idea of Taylor Remainder Theorem. I want to know the way I understand it is right or wrong.



    Like in the linear approximation



    it is $$f(x) approx f(c) + f'(c)(x-c)$$
    or $$f(x) = f(c) + f'(c)(x-c) + f''(zeta)(x-c)$$



    which is in the picture below
    enter image description here



    The error is exactly $$f''(zeta)(x-c)$$



    Then I thought of a case with a lower order of derivative which is



    $$f(x) = f(c) + f'(zeta)(x-c)$$



    enter image description here



    The error is exactly $$f'(zeta)(x-c)$$



    Then I noticed as there are more terms the order of derivatives for error gets higher which leads to the equation $$fracf^n+1(zeta)(n+1)!(x-a)^n+1$$



    Is this the right way of understanding the theorem?







    share|cite|improve this question
























      up vote
      0
      down vote

      favorite









      up vote
      0
      down vote

      favorite











      I am trying to grasp the idea of Taylor Remainder Theorem. I want to know the way I understand it is right or wrong.



      Like in the linear approximation



      it is $$f(x) approx f(c) + f'(c)(x-c)$$
      or $$f(x) = f(c) + f'(c)(x-c) + f''(zeta)(x-c)$$



      which is in the picture below
      enter image description here



      The error is exactly $$f''(zeta)(x-c)$$



      Then I thought of a case with a lower order of derivative which is



      $$f(x) = f(c) + f'(zeta)(x-c)$$



      enter image description here



      The error is exactly $$f'(zeta)(x-c)$$



      Then I noticed as there are more terms the order of derivatives for error gets higher which leads to the equation $$fracf^n+1(zeta)(n+1)!(x-a)^n+1$$



      Is this the right way of understanding the theorem?







      share|cite|improve this question














      I am trying to grasp the idea of Taylor Remainder Theorem. I want to know the way I understand it is right or wrong.



      Like in the linear approximation



      it is $$f(x) approx f(c) + f'(c)(x-c)$$
      or $$f(x) = f(c) + f'(c)(x-c) + f''(zeta)(x-c)$$



      which is in the picture below
      enter image description here



      The error is exactly $$f''(zeta)(x-c)$$



      Then I thought of a case with a lower order of derivative which is



      $$f(x) = f(c) + f'(zeta)(x-c)$$



      enter image description here



      The error is exactly $$f'(zeta)(x-c)$$



      Then I noticed as there are more terms the order of derivatives for error gets higher which leads to the equation $$fracf^n+1(zeta)(n+1)!(x-a)^n+1$$



      Is this the right way of understanding the theorem?









      share|cite|improve this question













      share|cite|improve this question




      share|cite|improve this question








      edited Aug 16 at 11:02









      Lukas Kofler

      6211518




      6211518










      asked Aug 16 at 10:47









      강승태

      173




      173




















          1 Answer
          1






          active

          oldest

          votes

















          up vote
          0
          down vote













          If $f$ is sufficiently often differentiable in the neighborhood of the point $cinmathbb R$ then for each $ngeq0$ its $n^rm th$ Taylor polynomial $j_c^nf$ is defined as follows:
          $$j_c^nf(x):=sum_k=0^nf^(k)(c)over k!(x-c)^k .$$
          $bigl($Note: Sometimes the increment variable $X:=x-c$ is used as variable for $j_c^nf$. One then writes
          $$j_c^nf(X):=sum_k=0^nf^(k)(c)over k!X^k .bigr)$$
          Given $x$ the value of such a polynomial can be computed exactly in finitely many steps.



          Why should we introduce this polynomial? That's where "Taylor's theorem with remainder" comes in. It turns out that when $|x-c|$ is small this polynomial gives a good approximation to the true value of $f$ at $x$:
          $$f(x)approx j_c^nf (x)quadbigl(|x-c|ll1) .$$
          Now "good approximation" is just a colloquial description. We want error bounds! There are various ways to quantify the error
          $$R_n(x):=f(x)-j_c^nf(x) .$$
          One of them reads as follows: There is a point $xi$ between $c$ and $x$ such that
          $$R_n(x)=f^(n+1)(xi)over (n+1)!(x-c)^n+1 .tag1$$
          The formula $(1)$ only is of use if you have simple control over the values of $f^(n+1)$. This is, e.g., the case when $f=sin$. But don't think the formula
          $$f(x)=j_c^nf(x)+R_n(x)$$
          allows you to compute the value of some special $f(x)$ exactly in a "finitary", albeit very complicated way.






          share|cite|improve this answer




















            Your Answer




            StackExchange.ifUsing("editor", function ()
            return StackExchange.using("mathjaxEditing", function ()
            StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix)
            StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
            );
            );
            , "mathjax-editing");

            StackExchange.ready(function()
            var channelOptions =
            tags: "".split(" "),
            id: "69"
            ;
            initTagRenderer("".split(" "), "".split(" "), channelOptions);

            StackExchange.using("externalEditor", function()
            // Have to fire editor after snippets, if snippets enabled
            if (StackExchange.settings.snippets.snippetsEnabled)
            StackExchange.using("snippets", function()
            createEditor();
            );

            else
            createEditor();

            );

            function createEditor()
            StackExchange.prepareEditor(
            heartbeatType: 'answer',
            convertImagesToLinks: true,
            noModals: false,
            showLowRepImageUploadWarning: true,
            reputationToPostImages: 10,
            bindNavPrevention: true,
            postfix: "",
            noCode: true, onDemand: true,
            discardSelector: ".discard-answer"
            ,immediatelyShowMarkdownHelp:true
            );



            );








             

            draft saved


            draft discarded


















            StackExchange.ready(
            function ()
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f2884651%2fgrasping-the-concept-of-taylor-remainder-theorem%23new-answer', 'question_page');

            );

            Post as a guest






























            1 Answer
            1






            active

            oldest

            votes








            1 Answer
            1






            active

            oldest

            votes









            active

            oldest

            votes






            active

            oldest

            votes








            up vote
            0
            down vote













            If $f$ is sufficiently often differentiable in the neighborhood of the point $cinmathbb R$ then for each $ngeq0$ its $n^rm th$ Taylor polynomial $j_c^nf$ is defined as follows:
            $$j_c^nf(x):=sum_k=0^nf^(k)(c)over k!(x-c)^k .$$
            $bigl($Note: Sometimes the increment variable $X:=x-c$ is used as variable for $j_c^nf$. One then writes
            $$j_c^nf(X):=sum_k=0^nf^(k)(c)over k!X^k .bigr)$$
            Given $x$ the value of such a polynomial can be computed exactly in finitely many steps.



            Why should we introduce this polynomial? That's where "Taylor's theorem with remainder" comes in. It turns out that when $|x-c|$ is small this polynomial gives a good approximation to the true value of $f$ at $x$:
            $$f(x)approx j_c^nf (x)quadbigl(|x-c|ll1) .$$
            Now "good approximation" is just a colloquial description. We want error bounds! There are various ways to quantify the error
            $$R_n(x):=f(x)-j_c^nf(x) .$$
            One of them reads as follows: There is a point $xi$ between $c$ and $x$ such that
            $$R_n(x)=f^(n+1)(xi)over (n+1)!(x-c)^n+1 .tag1$$
            The formula $(1)$ only is of use if you have simple control over the values of $f^(n+1)$. This is, e.g., the case when $f=sin$. But don't think the formula
            $$f(x)=j_c^nf(x)+R_n(x)$$
            allows you to compute the value of some special $f(x)$ exactly in a "finitary", albeit very complicated way.






            share|cite|improve this answer
























              up vote
              0
              down vote













              If $f$ is sufficiently often differentiable in the neighborhood of the point $cinmathbb R$ then for each $ngeq0$ its $n^rm th$ Taylor polynomial $j_c^nf$ is defined as follows:
              $$j_c^nf(x):=sum_k=0^nf^(k)(c)over k!(x-c)^k .$$
              $bigl($Note: Sometimes the increment variable $X:=x-c$ is used as variable for $j_c^nf$. One then writes
              $$j_c^nf(X):=sum_k=0^nf^(k)(c)over k!X^k .bigr)$$
              Given $x$ the value of such a polynomial can be computed exactly in finitely many steps.



              Why should we introduce this polynomial? That's where "Taylor's theorem with remainder" comes in. It turns out that when $|x-c|$ is small this polynomial gives a good approximation to the true value of $f$ at $x$:
              $$f(x)approx j_c^nf (x)quadbigl(|x-c|ll1) .$$
              Now "good approximation" is just a colloquial description. We want error bounds! There are various ways to quantify the error
              $$R_n(x):=f(x)-j_c^nf(x) .$$
              One of them reads as follows: There is a point $xi$ between $c$ and $x$ such that
              $$R_n(x)=f^(n+1)(xi)over (n+1)!(x-c)^n+1 .tag1$$
              The formula $(1)$ only is of use if you have simple control over the values of $f^(n+1)$. This is, e.g., the case when $f=sin$. But don't think the formula
              $$f(x)=j_c^nf(x)+R_n(x)$$
              allows you to compute the value of some special $f(x)$ exactly in a "finitary", albeit very complicated way.






              share|cite|improve this answer






















                up vote
                0
                down vote










                up vote
                0
                down vote









                If $f$ is sufficiently often differentiable in the neighborhood of the point $cinmathbb R$ then for each $ngeq0$ its $n^rm th$ Taylor polynomial $j_c^nf$ is defined as follows:
                $$j_c^nf(x):=sum_k=0^nf^(k)(c)over k!(x-c)^k .$$
                $bigl($Note: Sometimes the increment variable $X:=x-c$ is used as variable for $j_c^nf$. One then writes
                $$j_c^nf(X):=sum_k=0^nf^(k)(c)over k!X^k .bigr)$$
                Given $x$ the value of such a polynomial can be computed exactly in finitely many steps.



                Why should we introduce this polynomial? That's where "Taylor's theorem with remainder" comes in. It turns out that when $|x-c|$ is small this polynomial gives a good approximation to the true value of $f$ at $x$:
                $$f(x)approx j_c^nf (x)quadbigl(|x-c|ll1) .$$
                Now "good approximation" is just a colloquial description. We want error bounds! There are various ways to quantify the error
                $$R_n(x):=f(x)-j_c^nf(x) .$$
                One of them reads as follows: There is a point $xi$ between $c$ and $x$ such that
                $$R_n(x)=f^(n+1)(xi)over (n+1)!(x-c)^n+1 .tag1$$
                The formula $(1)$ only is of use if you have simple control over the values of $f^(n+1)$. This is, e.g., the case when $f=sin$. But don't think the formula
                $$f(x)=j_c^nf(x)+R_n(x)$$
                allows you to compute the value of some special $f(x)$ exactly in a "finitary", albeit very complicated way.






                share|cite|improve this answer












                If $f$ is sufficiently often differentiable in the neighborhood of the point $cinmathbb R$ then for each $ngeq0$ its $n^rm th$ Taylor polynomial $j_c^nf$ is defined as follows:
                $$j_c^nf(x):=sum_k=0^nf^(k)(c)over k!(x-c)^k .$$
                $bigl($Note: Sometimes the increment variable $X:=x-c$ is used as variable for $j_c^nf$. One then writes
                $$j_c^nf(X):=sum_k=0^nf^(k)(c)over k!X^k .bigr)$$
                Given $x$ the value of such a polynomial can be computed exactly in finitely many steps.



                Why should we introduce this polynomial? That's where "Taylor's theorem with remainder" comes in. It turns out that when $|x-c|$ is small this polynomial gives a good approximation to the true value of $f$ at $x$:
                $$f(x)approx j_c^nf (x)quadbigl(|x-c|ll1) .$$
                Now "good approximation" is just a colloquial description. We want error bounds! There are various ways to quantify the error
                $$R_n(x):=f(x)-j_c^nf(x) .$$
                One of them reads as follows: There is a point $xi$ between $c$ and $x$ such that
                $$R_n(x)=f^(n+1)(xi)over (n+1)!(x-c)^n+1 .tag1$$
                The formula $(1)$ only is of use if you have simple control over the values of $f^(n+1)$. This is, e.g., the case when $f=sin$. But don't think the formula
                $$f(x)=j_c^nf(x)+R_n(x)$$
                allows you to compute the value of some special $f(x)$ exactly in a "finitary", albeit very complicated way.







                share|cite|improve this answer












                share|cite|improve this answer



                share|cite|improve this answer










                answered Aug 16 at 14:09









                Christian Blatter

                165k7109309




                165k7109309






















                     

                    draft saved


                    draft discarded


























                     


                    draft saved


                    draft discarded














                    StackExchange.ready(
                    function ()
                    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f2884651%2fgrasping-the-concept-of-taylor-remainder-theorem%23new-answer', 'question_page');

                    );

                    Post as a guest













































































                    這個網誌中的熱門文章

                    Is there any way to eliminate the singular point to solve this integral by hand or by approximations?

                    Why am i infinitely getting the same tweet with the Twitter Search API?

                    Carbon dioxide