The reasoning behind doing series expansions and approximating functions in physics












12














It is usual in physics, that when we have a variable that is very small or very large we do a power series expansion of the function of that variable, and eliminate the high order terms, but my question is, why do we usually make the expansion and then approximate, why don't we just do the limit in that function, when that value is very small (tends to zero) or is very large (tends to infinity).










share|cite|improve this question





























    12














    It is usual in physics, that when we have a variable that is very small or very large we do a power series expansion of the function of that variable, and eliminate the high order terms, but my question is, why do we usually make the expansion and then approximate, why don't we just do the limit in that function, when that value is very small (tends to zero) or is very large (tends to infinity).










    share|cite|improve this question



























      12












      12








      12


      1





      It is usual in physics, that when we have a variable that is very small or very large we do a power series expansion of the function of that variable, and eliminate the high order terms, but my question is, why do we usually make the expansion and then approximate, why don't we just do the limit in that function, when that value is very small (tends to zero) or is very large (tends to infinity).










      share|cite|improve this question















      It is usual in physics, that when we have a variable that is very small or very large we do a power series expansion of the function of that variable, and eliminate the high order terms, but my question is, why do we usually make the expansion and then approximate, why don't we just do the limit in that function, when that value is very small (tends to zero) or is very large (tends to infinity).







      mathematics approximations






      share|cite|improve this question















      share|cite|improve this question













      share|cite|improve this question




      share|cite|improve this question








      edited yesterday









      Qmechanic

      101k121831152




      101k121831152










      asked yesterday









      orochi

      846




      846






















          4 Answers
          4






          active

          oldest

          votes


















          34














          The key reason is that we want to understand the behavior of the system in the neighborhood of the state rather than at the state itself.



          Take the equation of motion for a simple pendulum, for example:



          $$ddot{theta} = -frac{g}{ell}sin(theta)$$



          If we take the limit where $theta rightarrow 0$, we find $ddot{theta}= 0$, and we would conclude that the pendulum angle increases or decreases linearly with respect to time.



          If we however take a Taylor expansion and truncate at the linear term, we find $ddot{theta} = -frac{g}{ell}theta$, which is a simple harmonic oscillator! This expansion shows us that in the neighborhood of $0$, the system returns back to $0$ as if it was a simple harmonic oscillator: completely unlike what we could state in the limit approximation above.



          In fact, you could consider the limiting behavior around a state to be the zeroth-order component of a local expansion, which holds true straightforwardly for the example above since the limit term contributes no terms to the dynamics of the pendulum (but correctly notes that the angle increases/decreases linearly very close to $0$).






          share|cite|improve this answer





























            9














            The idea behind any expansion is to express a "complicated" function in terms of simpler ones. In the case of a series expansion, the simpler ones are polynomials. Thus for instance the function
            $$
            frac{1}{a+x}-frac{1}{a-x} tag{1}
            $$

            is a difference of two approximately equal quantities when $x/ato 0$ and so appears to be $0$ when $x/ato 0$ but that's not really useful information so it is convenient to reexpress it as
            $$
            frac{1}{a(1+x/a)}-frac{1}{a(1-x/a)} approx
            -frac{2 x}{a^2}-frac{2 x^3}{a^4}
            $$

            which gives some additional information in this limit.



            There are also multiple circumstances where some equations - say a differential equation - cannot be solved exactly but can be solved in some limit (often yielding a linearized equation or systems of equations), which still allows some qualitative understanding of the features of the solutions: this is the basis for perturbation theory. For instance, solving the Schrodinger equation for the Lennard-Jones potential
            $$
            V(r)= 4epsilonleft[left(frac{sigma}{r}right)^{12}
            -left(frac{sigma}{r}right)^6right]
            $$

            cannot be done analytically, but near the minimum $r_0=2^{1/6}sigma$ one can expand to obtain
            $$
            V(r)approx -epsilon+ frac{18 2^{2/3} epsilon (r-r_0)^2}{sigma ^2}
            $$

            which, up to an unimportant constant and shift in $r$, is a harmonic oscillator potential for which the solutions are known. Thus, we can get some approximate insight (or at least some orders of magnitude) into the appropriate molecular transitions.



            Of course great care must be taken to insure that the assumptions behind the expansion are not ultimately violated by the solutions, i.e. one must understand that the resulting solutions are approximate and may fail badly in some cases.






            share|cite|improve this answer































              6














              Consider the function $f(x)$ defined by
              $$
              f(x)equiv int^infty_{-infty} ds big(exp(-s^2-xs^4) - exp(-s^2)big).
              tag{1}
              $$

              When $x=0$, we get $f(0)=0$. What if we want to know the value of $f(x)$ when $x$ is a very small positive number? We don't know how to evaluate this integral exactly and explicitly, and just saying that the result will be "close to zero" is not very enlightening.



              We could evaluate the integral numerically, but that requires a computer (or a very patient person with a lot of time), and if we do the calculation that way, then we have to re-do it for each new value of $x$ that we care about.



              An alternative is to expand in powers of $x$:
              $$
              f(x)
              = int^infty_{-infty} ds (-xs^4)exp(-s^2)
              + int^infty_{-infty} ds frac{(-xs^4)^2}{2!}exp(-s^2)
              + int^infty_{-infty} ds frac{(-xs^4)^3}{3!}exp(-s^2)
              +cdots
              tag{2}
              $$

              Each term in this expansion is an elementary integral, which can be evaluated explicitly, so we end up a series in powers of $x$ with explicit numeric coefficients. The expansion doesn't converge (it's an asymptotic expansion), but if $x$ is small enough, then the first few terms give a good approximation, and we don't have to re-compute the coefficients every time we want to try a new value of $x$.



              Examples like this are everywhere in physics. This particular example is the single-variable version of an integral that shows up in the simplest type of non-trivial quantum field theory (the "$phi^4$ model").






              share|cite|improve this answer





























                0














                In general, one uses whatever works to learn something about the system.



                Get an exact exact solution if you can. But too often that is not possiblr.



                So, use any technique you like to learn something about the behaviour.



                Turns out that the perturbative expansion can often be used, is usually meaningful around a stable state and is helpful. So, it becomes a golden hammer.



                However, it is good to be skeptical as to validity in a given case. For systems in a state far away from a stable minimum such techniques are often not valid at all.






                share|cite|improve this answer





















                  Your Answer





                  StackExchange.ifUsing("editor", function () {
                  return StackExchange.using("mathjaxEditing", function () {
                  StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
                  StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
                  });
                  });
                  }, "mathjax-editing");

                  StackExchange.ready(function() {
                  var channelOptions = {
                  tags: "".split(" "),
                  id: "151"
                  };
                  initTagRenderer("".split(" "), "".split(" "), channelOptions);

                  StackExchange.using("externalEditor", function() {
                  // Have to fire editor after snippets, if snippets enabled
                  if (StackExchange.settings.snippets.snippetsEnabled) {
                  StackExchange.using("snippets", function() {
                  createEditor();
                  });
                  }
                  else {
                  createEditor();
                  }
                  });

                  function createEditor() {
                  StackExchange.prepareEditor({
                  heartbeatType: 'answer',
                  autoActivateHeartbeat: false,
                  convertImagesToLinks: false,
                  noModals: true,
                  showLowRepImageUploadWarning: true,
                  reputationToPostImages: null,
                  bindNavPrevention: true,
                  postfix: "",
                  imageUploader: {
                  brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
                  contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
                  allowUrls: true
                  },
                  noCode: true, onDemand: true,
                  discardSelector: ".discard-answer"
                  ,immediatelyShowMarkdownHelp:true
                  });


                  }
                  });














                  draft saved

                  draft discarded


















                  StackExchange.ready(
                  function () {
                  StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fphysics.stackexchange.com%2fquestions%2f451588%2fthe-reasoning-behind-doing-series-expansions-and-approximating-functions-in-phys%23new-answer', 'question_page');
                  }
                  );

                  Post as a guest















                  Required, but never shown

























                  4 Answers
                  4






                  active

                  oldest

                  votes








                  4 Answers
                  4






                  active

                  oldest

                  votes









                  active

                  oldest

                  votes






                  active

                  oldest

                  votes









                  34














                  The key reason is that we want to understand the behavior of the system in the neighborhood of the state rather than at the state itself.



                  Take the equation of motion for a simple pendulum, for example:



                  $$ddot{theta} = -frac{g}{ell}sin(theta)$$



                  If we take the limit where $theta rightarrow 0$, we find $ddot{theta}= 0$, and we would conclude that the pendulum angle increases or decreases linearly with respect to time.



                  If we however take a Taylor expansion and truncate at the linear term, we find $ddot{theta} = -frac{g}{ell}theta$, which is a simple harmonic oscillator! This expansion shows us that in the neighborhood of $0$, the system returns back to $0$ as if it was a simple harmonic oscillator: completely unlike what we could state in the limit approximation above.



                  In fact, you could consider the limiting behavior around a state to be the zeroth-order component of a local expansion, which holds true straightforwardly for the example above since the limit term contributes no terms to the dynamics of the pendulum (but correctly notes that the angle increases/decreases linearly very close to $0$).






                  share|cite|improve this answer


























                    34














                    The key reason is that we want to understand the behavior of the system in the neighborhood of the state rather than at the state itself.



                    Take the equation of motion for a simple pendulum, for example:



                    $$ddot{theta} = -frac{g}{ell}sin(theta)$$



                    If we take the limit where $theta rightarrow 0$, we find $ddot{theta}= 0$, and we would conclude that the pendulum angle increases or decreases linearly with respect to time.



                    If we however take a Taylor expansion and truncate at the linear term, we find $ddot{theta} = -frac{g}{ell}theta$, which is a simple harmonic oscillator! This expansion shows us that in the neighborhood of $0$, the system returns back to $0$ as if it was a simple harmonic oscillator: completely unlike what we could state in the limit approximation above.



                    In fact, you could consider the limiting behavior around a state to be the zeroth-order component of a local expansion, which holds true straightforwardly for the example above since the limit term contributes no terms to the dynamics of the pendulum (but correctly notes that the angle increases/decreases linearly very close to $0$).






                    share|cite|improve this answer
























                      34












                      34








                      34






                      The key reason is that we want to understand the behavior of the system in the neighborhood of the state rather than at the state itself.



                      Take the equation of motion for a simple pendulum, for example:



                      $$ddot{theta} = -frac{g}{ell}sin(theta)$$



                      If we take the limit where $theta rightarrow 0$, we find $ddot{theta}= 0$, and we would conclude that the pendulum angle increases or decreases linearly with respect to time.



                      If we however take a Taylor expansion and truncate at the linear term, we find $ddot{theta} = -frac{g}{ell}theta$, which is a simple harmonic oscillator! This expansion shows us that in the neighborhood of $0$, the system returns back to $0$ as if it was a simple harmonic oscillator: completely unlike what we could state in the limit approximation above.



                      In fact, you could consider the limiting behavior around a state to be the zeroth-order component of a local expansion, which holds true straightforwardly for the example above since the limit term contributes no terms to the dynamics of the pendulum (but correctly notes that the angle increases/decreases linearly very close to $0$).






                      share|cite|improve this answer












                      The key reason is that we want to understand the behavior of the system in the neighborhood of the state rather than at the state itself.



                      Take the equation of motion for a simple pendulum, for example:



                      $$ddot{theta} = -frac{g}{ell}sin(theta)$$



                      If we take the limit where $theta rightarrow 0$, we find $ddot{theta}= 0$, and we would conclude that the pendulum angle increases or decreases linearly with respect to time.



                      If we however take a Taylor expansion and truncate at the linear term, we find $ddot{theta} = -frac{g}{ell}theta$, which is a simple harmonic oscillator! This expansion shows us that in the neighborhood of $0$, the system returns back to $0$ as if it was a simple harmonic oscillator: completely unlike what we could state in the limit approximation above.



                      In fact, you could consider the limiting behavior around a state to be the zeroth-order component of a local expansion, which holds true straightforwardly for the example above since the limit term contributes no terms to the dynamics of the pendulum (but correctly notes that the angle increases/decreases linearly very close to $0$).







                      share|cite|improve this answer












                      share|cite|improve this answer



                      share|cite|improve this answer










                      answered yesterday









                      aghostinthefigures

                      1,018217




                      1,018217























                          9














                          The idea behind any expansion is to express a "complicated" function in terms of simpler ones. In the case of a series expansion, the simpler ones are polynomials. Thus for instance the function
                          $$
                          frac{1}{a+x}-frac{1}{a-x} tag{1}
                          $$

                          is a difference of two approximately equal quantities when $x/ato 0$ and so appears to be $0$ when $x/ato 0$ but that's not really useful information so it is convenient to reexpress it as
                          $$
                          frac{1}{a(1+x/a)}-frac{1}{a(1-x/a)} approx
                          -frac{2 x}{a^2}-frac{2 x^3}{a^4}
                          $$

                          which gives some additional information in this limit.



                          There are also multiple circumstances where some equations - say a differential equation - cannot be solved exactly but can be solved in some limit (often yielding a linearized equation or systems of equations), which still allows some qualitative understanding of the features of the solutions: this is the basis for perturbation theory. For instance, solving the Schrodinger equation for the Lennard-Jones potential
                          $$
                          V(r)= 4epsilonleft[left(frac{sigma}{r}right)^{12}
                          -left(frac{sigma}{r}right)^6right]
                          $$

                          cannot be done analytically, but near the minimum $r_0=2^{1/6}sigma$ one can expand to obtain
                          $$
                          V(r)approx -epsilon+ frac{18 2^{2/3} epsilon (r-r_0)^2}{sigma ^2}
                          $$

                          which, up to an unimportant constant and shift in $r$, is a harmonic oscillator potential for which the solutions are known. Thus, we can get some approximate insight (or at least some orders of magnitude) into the appropriate molecular transitions.



                          Of course great care must be taken to insure that the assumptions behind the expansion are not ultimately violated by the solutions, i.e. one must understand that the resulting solutions are approximate and may fail badly in some cases.






                          share|cite|improve this answer




























                            9














                            The idea behind any expansion is to express a "complicated" function in terms of simpler ones. In the case of a series expansion, the simpler ones are polynomials. Thus for instance the function
                            $$
                            frac{1}{a+x}-frac{1}{a-x} tag{1}
                            $$

                            is a difference of two approximately equal quantities when $x/ato 0$ and so appears to be $0$ when $x/ato 0$ but that's not really useful information so it is convenient to reexpress it as
                            $$
                            frac{1}{a(1+x/a)}-frac{1}{a(1-x/a)} approx
                            -frac{2 x}{a^2}-frac{2 x^3}{a^4}
                            $$

                            which gives some additional information in this limit.



                            There are also multiple circumstances where some equations - say a differential equation - cannot be solved exactly but can be solved in some limit (often yielding a linearized equation or systems of equations), which still allows some qualitative understanding of the features of the solutions: this is the basis for perturbation theory. For instance, solving the Schrodinger equation for the Lennard-Jones potential
                            $$
                            V(r)= 4epsilonleft[left(frac{sigma}{r}right)^{12}
                            -left(frac{sigma}{r}right)^6right]
                            $$

                            cannot be done analytically, but near the minimum $r_0=2^{1/6}sigma$ one can expand to obtain
                            $$
                            V(r)approx -epsilon+ frac{18 2^{2/3} epsilon (r-r_0)^2}{sigma ^2}
                            $$

                            which, up to an unimportant constant and shift in $r$, is a harmonic oscillator potential for which the solutions are known. Thus, we can get some approximate insight (or at least some orders of magnitude) into the appropriate molecular transitions.



                            Of course great care must be taken to insure that the assumptions behind the expansion are not ultimately violated by the solutions, i.e. one must understand that the resulting solutions are approximate and may fail badly in some cases.






                            share|cite|improve this answer


























                              9












                              9








                              9






                              The idea behind any expansion is to express a "complicated" function in terms of simpler ones. In the case of a series expansion, the simpler ones are polynomials. Thus for instance the function
                              $$
                              frac{1}{a+x}-frac{1}{a-x} tag{1}
                              $$

                              is a difference of two approximately equal quantities when $x/ato 0$ and so appears to be $0$ when $x/ato 0$ but that's not really useful information so it is convenient to reexpress it as
                              $$
                              frac{1}{a(1+x/a)}-frac{1}{a(1-x/a)} approx
                              -frac{2 x}{a^2}-frac{2 x^3}{a^4}
                              $$

                              which gives some additional information in this limit.



                              There are also multiple circumstances where some equations - say a differential equation - cannot be solved exactly but can be solved in some limit (often yielding a linearized equation or systems of equations), which still allows some qualitative understanding of the features of the solutions: this is the basis for perturbation theory. For instance, solving the Schrodinger equation for the Lennard-Jones potential
                              $$
                              V(r)= 4epsilonleft[left(frac{sigma}{r}right)^{12}
                              -left(frac{sigma}{r}right)^6right]
                              $$

                              cannot be done analytically, but near the minimum $r_0=2^{1/6}sigma$ one can expand to obtain
                              $$
                              V(r)approx -epsilon+ frac{18 2^{2/3} epsilon (r-r_0)^2}{sigma ^2}
                              $$

                              which, up to an unimportant constant and shift in $r$, is a harmonic oscillator potential for which the solutions are known. Thus, we can get some approximate insight (or at least some orders of magnitude) into the appropriate molecular transitions.



                              Of course great care must be taken to insure that the assumptions behind the expansion are not ultimately violated by the solutions, i.e. one must understand that the resulting solutions are approximate and may fail badly in some cases.






                              share|cite|improve this answer














                              The idea behind any expansion is to express a "complicated" function in terms of simpler ones. In the case of a series expansion, the simpler ones are polynomials. Thus for instance the function
                              $$
                              frac{1}{a+x}-frac{1}{a-x} tag{1}
                              $$

                              is a difference of two approximately equal quantities when $x/ato 0$ and so appears to be $0$ when $x/ato 0$ but that's not really useful information so it is convenient to reexpress it as
                              $$
                              frac{1}{a(1+x/a)}-frac{1}{a(1-x/a)} approx
                              -frac{2 x}{a^2}-frac{2 x^3}{a^4}
                              $$

                              which gives some additional information in this limit.



                              There are also multiple circumstances where some equations - say a differential equation - cannot be solved exactly but can be solved in some limit (often yielding a linearized equation or systems of equations), which still allows some qualitative understanding of the features of the solutions: this is the basis for perturbation theory. For instance, solving the Schrodinger equation for the Lennard-Jones potential
                              $$
                              V(r)= 4epsilonleft[left(frac{sigma}{r}right)^{12}
                              -left(frac{sigma}{r}right)^6right]
                              $$

                              cannot be done analytically, but near the minimum $r_0=2^{1/6}sigma$ one can expand to obtain
                              $$
                              V(r)approx -epsilon+ frac{18 2^{2/3} epsilon (r-r_0)^2}{sigma ^2}
                              $$

                              which, up to an unimportant constant and shift in $r$, is a harmonic oscillator potential for which the solutions are known. Thus, we can get some approximate insight (or at least some orders of magnitude) into the appropriate molecular transitions.



                              Of course great care must be taken to insure that the assumptions behind the expansion are not ultimately violated by the solutions, i.e. one must understand that the resulting solutions are approximate and may fail badly in some cases.







                              share|cite|improve this answer














                              share|cite|improve this answer



                              share|cite|improve this answer








                              edited yesterday

























                              answered yesterday









                              ZeroTheHero

                              18.7k52956




                              18.7k52956























                                  6














                                  Consider the function $f(x)$ defined by
                                  $$
                                  f(x)equiv int^infty_{-infty} ds big(exp(-s^2-xs^4) - exp(-s^2)big).
                                  tag{1}
                                  $$

                                  When $x=0$, we get $f(0)=0$. What if we want to know the value of $f(x)$ when $x$ is a very small positive number? We don't know how to evaluate this integral exactly and explicitly, and just saying that the result will be "close to zero" is not very enlightening.



                                  We could evaluate the integral numerically, but that requires a computer (or a very patient person with a lot of time), and if we do the calculation that way, then we have to re-do it for each new value of $x$ that we care about.



                                  An alternative is to expand in powers of $x$:
                                  $$
                                  f(x)
                                  = int^infty_{-infty} ds (-xs^4)exp(-s^2)
                                  + int^infty_{-infty} ds frac{(-xs^4)^2}{2!}exp(-s^2)
                                  + int^infty_{-infty} ds frac{(-xs^4)^3}{3!}exp(-s^2)
                                  +cdots
                                  tag{2}
                                  $$

                                  Each term in this expansion is an elementary integral, which can be evaluated explicitly, so we end up a series in powers of $x$ with explicit numeric coefficients. The expansion doesn't converge (it's an asymptotic expansion), but if $x$ is small enough, then the first few terms give a good approximation, and we don't have to re-compute the coefficients every time we want to try a new value of $x$.



                                  Examples like this are everywhere in physics. This particular example is the single-variable version of an integral that shows up in the simplest type of non-trivial quantum field theory (the "$phi^4$ model").






                                  share|cite|improve this answer


























                                    6














                                    Consider the function $f(x)$ defined by
                                    $$
                                    f(x)equiv int^infty_{-infty} ds big(exp(-s^2-xs^4) - exp(-s^2)big).
                                    tag{1}
                                    $$

                                    When $x=0$, we get $f(0)=0$. What if we want to know the value of $f(x)$ when $x$ is a very small positive number? We don't know how to evaluate this integral exactly and explicitly, and just saying that the result will be "close to zero" is not very enlightening.



                                    We could evaluate the integral numerically, but that requires a computer (or a very patient person with a lot of time), and if we do the calculation that way, then we have to re-do it for each new value of $x$ that we care about.



                                    An alternative is to expand in powers of $x$:
                                    $$
                                    f(x)
                                    = int^infty_{-infty} ds (-xs^4)exp(-s^2)
                                    + int^infty_{-infty} ds frac{(-xs^4)^2}{2!}exp(-s^2)
                                    + int^infty_{-infty} ds frac{(-xs^4)^3}{3!}exp(-s^2)
                                    +cdots
                                    tag{2}
                                    $$

                                    Each term in this expansion is an elementary integral, which can be evaluated explicitly, so we end up a series in powers of $x$ with explicit numeric coefficients. The expansion doesn't converge (it's an asymptotic expansion), but if $x$ is small enough, then the first few terms give a good approximation, and we don't have to re-compute the coefficients every time we want to try a new value of $x$.



                                    Examples like this are everywhere in physics. This particular example is the single-variable version of an integral that shows up in the simplest type of non-trivial quantum field theory (the "$phi^4$ model").






                                    share|cite|improve this answer
























                                      6












                                      6








                                      6






                                      Consider the function $f(x)$ defined by
                                      $$
                                      f(x)equiv int^infty_{-infty} ds big(exp(-s^2-xs^4) - exp(-s^2)big).
                                      tag{1}
                                      $$

                                      When $x=0$, we get $f(0)=0$. What if we want to know the value of $f(x)$ when $x$ is a very small positive number? We don't know how to evaluate this integral exactly and explicitly, and just saying that the result will be "close to zero" is not very enlightening.



                                      We could evaluate the integral numerically, but that requires a computer (or a very patient person with a lot of time), and if we do the calculation that way, then we have to re-do it for each new value of $x$ that we care about.



                                      An alternative is to expand in powers of $x$:
                                      $$
                                      f(x)
                                      = int^infty_{-infty} ds (-xs^4)exp(-s^2)
                                      + int^infty_{-infty} ds frac{(-xs^4)^2}{2!}exp(-s^2)
                                      + int^infty_{-infty} ds frac{(-xs^4)^3}{3!}exp(-s^2)
                                      +cdots
                                      tag{2}
                                      $$

                                      Each term in this expansion is an elementary integral, which can be evaluated explicitly, so we end up a series in powers of $x$ with explicit numeric coefficients. The expansion doesn't converge (it's an asymptotic expansion), but if $x$ is small enough, then the first few terms give a good approximation, and we don't have to re-compute the coefficients every time we want to try a new value of $x$.



                                      Examples like this are everywhere in physics. This particular example is the single-variable version of an integral that shows up in the simplest type of non-trivial quantum field theory (the "$phi^4$ model").






                                      share|cite|improve this answer












                                      Consider the function $f(x)$ defined by
                                      $$
                                      f(x)equiv int^infty_{-infty} ds big(exp(-s^2-xs^4) - exp(-s^2)big).
                                      tag{1}
                                      $$

                                      When $x=0$, we get $f(0)=0$. What if we want to know the value of $f(x)$ when $x$ is a very small positive number? We don't know how to evaluate this integral exactly and explicitly, and just saying that the result will be "close to zero" is not very enlightening.



                                      We could evaluate the integral numerically, but that requires a computer (or a very patient person with a lot of time), and if we do the calculation that way, then we have to re-do it for each new value of $x$ that we care about.



                                      An alternative is to expand in powers of $x$:
                                      $$
                                      f(x)
                                      = int^infty_{-infty} ds (-xs^4)exp(-s^2)
                                      + int^infty_{-infty} ds frac{(-xs^4)^2}{2!}exp(-s^2)
                                      + int^infty_{-infty} ds frac{(-xs^4)^3}{3!}exp(-s^2)
                                      +cdots
                                      tag{2}
                                      $$

                                      Each term in this expansion is an elementary integral, which can be evaluated explicitly, so we end up a series in powers of $x$ with explicit numeric coefficients. The expansion doesn't converge (it's an asymptotic expansion), but if $x$ is small enough, then the first few terms give a good approximation, and we don't have to re-compute the coefficients every time we want to try a new value of $x$.



                                      Examples like this are everywhere in physics. This particular example is the single-variable version of an integral that shows up in the simplest type of non-trivial quantum field theory (the "$phi^4$ model").







                                      share|cite|improve this answer












                                      share|cite|improve this answer



                                      share|cite|improve this answer










                                      answered yesterday









                                      Dan Yand

                                      7,1321930




                                      7,1321930























                                          0














                                          In general, one uses whatever works to learn something about the system.



                                          Get an exact exact solution if you can. But too often that is not possiblr.



                                          So, use any technique you like to learn something about the behaviour.



                                          Turns out that the perturbative expansion can often be used, is usually meaningful around a stable state and is helpful. So, it becomes a golden hammer.



                                          However, it is good to be skeptical as to validity in a given case. For systems in a state far away from a stable minimum such techniques are often not valid at all.






                                          share|cite|improve this answer


























                                            0














                                            In general, one uses whatever works to learn something about the system.



                                            Get an exact exact solution if you can. But too often that is not possiblr.



                                            So, use any technique you like to learn something about the behaviour.



                                            Turns out that the perturbative expansion can often be used, is usually meaningful around a stable state and is helpful. So, it becomes a golden hammer.



                                            However, it is good to be skeptical as to validity in a given case. For systems in a state far away from a stable minimum such techniques are often not valid at all.






                                            share|cite|improve this answer
























                                              0












                                              0








                                              0






                                              In general, one uses whatever works to learn something about the system.



                                              Get an exact exact solution if you can. But too often that is not possiblr.



                                              So, use any technique you like to learn something about the behaviour.



                                              Turns out that the perturbative expansion can often be used, is usually meaningful around a stable state and is helpful. So, it becomes a golden hammer.



                                              However, it is good to be skeptical as to validity in a given case. For systems in a state far away from a stable minimum such techniques are often not valid at all.






                                              share|cite|improve this answer












                                              In general, one uses whatever works to learn something about the system.



                                              Get an exact exact solution if you can. But too often that is not possiblr.



                                              So, use any technique you like to learn something about the behaviour.



                                              Turns out that the perturbative expansion can often be used, is usually meaningful around a stable state and is helpful. So, it becomes a golden hammer.



                                              However, it is good to be skeptical as to validity in a given case. For systems in a state far away from a stable minimum such techniques are often not valid at all.







                                              share|cite|improve this answer












                                              share|cite|improve this answer



                                              share|cite|improve this answer










                                              answered 18 hours ago









                                              Keith

                                              54026




                                              54026






























                                                  draft saved

                                                  draft discarded




















































                                                  Thanks for contributing an answer to Physics Stack Exchange!


                                                  • Please be sure to answer the question. Provide details and share your research!

                                                  But avoid



                                                  • Asking for help, clarification, or responding to other answers.

                                                  • Making statements based on opinion; back them up with references or personal experience.


                                                  Use MathJax to format equations. MathJax reference.


                                                  To learn more, see our tips on writing great answers.





                                                  Some of your past answers have not been well-received, and you're in danger of being blocked from answering.


                                                  Please pay close attention to the following guidance:


                                                  • Please be sure to answer the question. Provide details and share your research!

                                                  But avoid



                                                  • Asking for help, clarification, or responding to other answers.

                                                  • Making statements based on opinion; back them up with references or personal experience.


                                                  To learn more, see our tips on writing great answers.




                                                  draft saved


                                                  draft discarded














                                                  StackExchange.ready(
                                                  function () {
                                                  StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fphysics.stackexchange.com%2fquestions%2f451588%2fthe-reasoning-behind-doing-series-expansions-and-approximating-functions-in-phys%23new-answer', 'question_page');
                                                  }
                                                  );

                                                  Post as a guest















                                                  Required, but never shown





















































                                                  Required, but never shown














                                                  Required, but never shown












                                                  Required, but never shown







                                                  Required, but never shown

































                                                  Required, but never shown














                                                  Required, but never shown












                                                  Required, but never shown







                                                  Required, but never shown







                                                  Popular posts from this blog

                                                  An IMO inspired problem

                                                  Management

                                                  Investment