Does convergence in $D'(Omega)$ imply convergence in $L^2(Omega)$?

The name of the pictureThe name of the pictureThe name of the pictureClash Royale CLAN TAG#URR8PPP











up vote
1
down vote

favorite












Consider $Omega subset mathbb R^n$ open.



Let $f_n_ngeq 1$ and $f$ belong to $L^2(Omega)$. Suppose that $f_n xrightarrowntoinfty f$ in $D'(Omega)$, i.e., $int_Omega f_n varphi xrightarrowntoinfty int_Omega fvarphi$, for all $varphi in C_0^infty(Omega)$. Are we able to conclude that $f_n xrightarrowntoinfty f$ in $L^2(Omega)$?



I think that, since $D(Omega)$ is dense in $L^2(Omega)$, we can conclude that $int_Omega f_n v xrightarrowntoinfty int_Omega f v $, for all $v in L^2(Omega)$ (am I right?). Then, maybe we can use Riesz representation theorem to write
$$
||f_n-f||_0,Omega= sup_substackv in L^2(Omega)\vneq 0 fraclangle f_n-f,vrangle_0,Omega,
$$
and somehow take the limit $ntoinfty$ inside the supremum.



EDIT:



Kavi Rama Murthy has shown that this is false if $Omega$ is unbounded. Is it different if $Omega$ is bounded?










share|cite|improve this question























  • What is $D'(Omega)$? What assumptions are there on the structure of $Omega$?
    – Stella Biderman
    Aug 30 at 4:16














up vote
1
down vote

favorite












Consider $Omega subset mathbb R^n$ open.



Let $f_n_ngeq 1$ and $f$ belong to $L^2(Omega)$. Suppose that $f_n xrightarrowntoinfty f$ in $D'(Omega)$, i.e., $int_Omega f_n varphi xrightarrowntoinfty int_Omega fvarphi$, for all $varphi in C_0^infty(Omega)$. Are we able to conclude that $f_n xrightarrowntoinfty f$ in $L^2(Omega)$?



I think that, since $D(Omega)$ is dense in $L^2(Omega)$, we can conclude that $int_Omega f_n v xrightarrowntoinfty int_Omega f v $, for all $v in L^2(Omega)$ (am I right?). Then, maybe we can use Riesz representation theorem to write
$$
||f_n-f||_0,Omega= sup_substackv in L^2(Omega)\vneq 0 fraclangle f_n-f,vrangle_0,Omega,
$$
and somehow take the limit $ntoinfty$ inside the supremum.



EDIT:



Kavi Rama Murthy has shown that this is false if $Omega$ is unbounded. Is it different if $Omega$ is bounded?










share|cite|improve this question























  • What is $D'(Omega)$? What assumptions are there on the structure of $Omega$?
    – Stella Biderman
    Aug 30 at 4:16












up vote
1
down vote

favorite









up vote
1
down vote

favorite











Consider $Omega subset mathbb R^n$ open.



Let $f_n_ngeq 1$ and $f$ belong to $L^2(Omega)$. Suppose that $f_n xrightarrowntoinfty f$ in $D'(Omega)$, i.e., $int_Omega f_n varphi xrightarrowntoinfty int_Omega fvarphi$, for all $varphi in C_0^infty(Omega)$. Are we able to conclude that $f_n xrightarrowntoinfty f$ in $L^2(Omega)$?



I think that, since $D(Omega)$ is dense in $L^2(Omega)$, we can conclude that $int_Omega f_n v xrightarrowntoinfty int_Omega f v $, for all $v in L^2(Omega)$ (am I right?). Then, maybe we can use Riesz representation theorem to write
$$
||f_n-f||_0,Omega= sup_substackv in L^2(Omega)\vneq 0 fraclangle f_n-f,vrangle_0,Omega,
$$
and somehow take the limit $ntoinfty$ inside the supremum.



EDIT:



Kavi Rama Murthy has shown that this is false if $Omega$ is unbounded. Is it different if $Omega$ is bounded?










share|cite|improve this question















Consider $Omega subset mathbb R^n$ open.



Let $f_n_ngeq 1$ and $f$ belong to $L^2(Omega)$. Suppose that $f_n xrightarrowntoinfty f$ in $D'(Omega)$, i.e., $int_Omega f_n varphi xrightarrowntoinfty int_Omega fvarphi$, for all $varphi in C_0^infty(Omega)$. Are we able to conclude that $f_n xrightarrowntoinfty f$ in $L^2(Omega)$?



I think that, since $D(Omega)$ is dense in $L^2(Omega)$, we can conclude that $int_Omega f_n v xrightarrowntoinfty int_Omega f v $, for all $v in L^2(Omega)$ (am I right?). Then, maybe we can use Riesz representation theorem to write
$$
||f_n-f||_0,Omega= sup_substackv in L^2(Omega)\vneq 0 fraclangle f_n-f,vrangle_0,Omega,
$$
and somehow take the limit $ntoinfty$ inside the supremum.



EDIT:



Kavi Rama Murthy has shown that this is false if $Omega$ is unbounded. Is it different if $Omega$ is bounded?







functional-analysis distribution-theory






share|cite|improve this question















share|cite|improve this question













share|cite|improve this question




share|cite|improve this question








edited Aug 30 at 21:29

























asked Aug 30 at 4:00









Gonzalo Benavides

628317




628317











  • What is $D'(Omega)$? What assumptions are there on the structure of $Omega$?
    – Stella Biderman
    Aug 30 at 4:16
















  • What is $D'(Omega)$? What assumptions are there on the structure of $Omega$?
    – Stella Biderman
    Aug 30 at 4:16















What is $D'(Omega)$? What assumptions are there on the structure of $Omega$?
– Stella Biderman
Aug 30 at 4:16




What is $D'(Omega)$? What assumptions are there on the structure of $Omega$?
– Stella Biderman
Aug 30 at 4:16










3 Answers
3






active

oldest

votes

















up vote
1
down vote













This is false. Let $f_n =I_(n,n+1)$, $f=0$. Then $int f_n phi to 0$ for all $phi in C_0^infty$ but $int |f_n-f|^2=1$ for all $n$.






share|cite|improve this answer




















  • Thanks! What if $Omega$ is bounded?
    – Gonzalo Benavides
    Aug 30 at 12:43










  • Then use vertical escape to $infty$ instead of horizontal escape to $infty$ to see that it's still false.
    – User8128
    Aug 31 at 2:26

















up vote
0
down vote













Suppose $Omega = (0,2pi).$ Let $f_n(x) = sin (nx).$ Then for any $varphi in C_c^infty(0,2pi),$ $int_0^2pi f_n varphi to 0$ by Riemann-Lebesgue. Thus $f_nto 0$ in $D'(0,2pi),$ but certainly not in $L^2(02pi).$






share|cite|improve this answer



























    up vote
    0
    down vote













    Another way to see this is the following:



    It is known that $D(Omega)$ is dense in $D'(Omega)$ and $L^2(Omega)$ with respect to their natural topologies. By considering delta functions, evidently $L^2(Omega) neq D'(Omega)$ if $Omega subset mathbb R^n$ is open.



    So if we take $u in D'(Omega) setminus L^2(Omega)$ and choose $u_n in D(Omega)$ such that $u_n rightarrow u$ in $D'(Omega),$ I claim $(u_n)$ does not converge in $L^2(Omega).$ Indeed if $u_n rightarrow v$ in $L^2(Omega),$ it converges weakly in particular and hence $u_n rightarrow v in D'(Omega).$ So $u = v in L^2(Omega),$ which is a contradiction.






    share|cite|improve this answer




















      Your Answer




      StackExchange.ifUsing("editor", function ()
      return StackExchange.using("mathjaxEditing", function ()
      StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix)
      StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
      );
      );
      , "mathjax-editing");

      StackExchange.ready(function()
      var channelOptions =
      tags: "".split(" "),
      id: "69"
      ;
      initTagRenderer("".split(" "), "".split(" "), channelOptions);

      StackExchange.using("externalEditor", function()
      // Have to fire editor after snippets, if snippets enabled
      if (StackExchange.settings.snippets.snippetsEnabled)
      StackExchange.using("snippets", function()
      createEditor();
      );

      else
      createEditor();

      );

      function createEditor()
      StackExchange.prepareEditor(
      heartbeatType: 'answer',
      convertImagesToLinks: true,
      noModals: false,
      showLowRepImageUploadWarning: true,
      reputationToPostImages: 10,
      bindNavPrevention: true,
      postfix: "",
      noCode: true, onDemand: true,
      discardSelector: ".discard-answer"
      ,immediatelyShowMarkdownHelp:true
      );



      );













       

      draft saved


      draft discarded


















      StackExchange.ready(
      function ()
      StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f2899105%2fdoes-convergence-in-d-omega-imply-convergence-in-l2-omega%23new-answer', 'question_page');

      );

      Post as a guest






























      3 Answers
      3






      active

      oldest

      votes








      3 Answers
      3






      active

      oldest

      votes









      active

      oldest

      votes






      active

      oldest

      votes








      up vote
      1
      down vote













      This is false. Let $f_n =I_(n,n+1)$, $f=0$. Then $int f_n phi to 0$ for all $phi in C_0^infty$ but $int |f_n-f|^2=1$ for all $n$.






      share|cite|improve this answer




















      • Thanks! What if $Omega$ is bounded?
        – Gonzalo Benavides
        Aug 30 at 12:43










      • Then use vertical escape to $infty$ instead of horizontal escape to $infty$ to see that it's still false.
        – User8128
        Aug 31 at 2:26














      up vote
      1
      down vote













      This is false. Let $f_n =I_(n,n+1)$, $f=0$. Then $int f_n phi to 0$ for all $phi in C_0^infty$ but $int |f_n-f|^2=1$ for all $n$.






      share|cite|improve this answer




















      • Thanks! What if $Omega$ is bounded?
        – Gonzalo Benavides
        Aug 30 at 12:43










      • Then use vertical escape to $infty$ instead of horizontal escape to $infty$ to see that it's still false.
        – User8128
        Aug 31 at 2:26












      up vote
      1
      down vote










      up vote
      1
      down vote









      This is false. Let $f_n =I_(n,n+1)$, $f=0$. Then $int f_n phi to 0$ for all $phi in C_0^infty$ but $int |f_n-f|^2=1$ for all $n$.






      share|cite|improve this answer












      This is false. Let $f_n =I_(n,n+1)$, $f=0$. Then $int f_n phi to 0$ for all $phi in C_0^infty$ but $int |f_n-f|^2=1$ for all $n$.







      share|cite|improve this answer












      share|cite|improve this answer



      share|cite|improve this answer










      answered Aug 30 at 6:04









      Kavi Rama Murthy

      25.3k31335




      25.3k31335











      • Thanks! What if $Omega$ is bounded?
        – Gonzalo Benavides
        Aug 30 at 12:43










      • Then use vertical escape to $infty$ instead of horizontal escape to $infty$ to see that it's still false.
        – User8128
        Aug 31 at 2:26
















      • Thanks! What if $Omega$ is bounded?
        – Gonzalo Benavides
        Aug 30 at 12:43










      • Then use vertical escape to $infty$ instead of horizontal escape to $infty$ to see that it's still false.
        – User8128
        Aug 31 at 2:26















      Thanks! What if $Omega$ is bounded?
      – Gonzalo Benavides
      Aug 30 at 12:43




      Thanks! What if $Omega$ is bounded?
      – Gonzalo Benavides
      Aug 30 at 12:43












      Then use vertical escape to $infty$ instead of horizontal escape to $infty$ to see that it's still false.
      – User8128
      Aug 31 at 2:26




      Then use vertical escape to $infty$ instead of horizontal escape to $infty$ to see that it's still false.
      – User8128
      Aug 31 at 2:26










      up vote
      0
      down vote













      Suppose $Omega = (0,2pi).$ Let $f_n(x) = sin (nx).$ Then for any $varphi in C_c^infty(0,2pi),$ $int_0^2pi f_n varphi to 0$ by Riemann-Lebesgue. Thus $f_nto 0$ in $D'(0,2pi),$ but certainly not in $L^2(02pi).$






      share|cite|improve this answer
























        up vote
        0
        down vote













        Suppose $Omega = (0,2pi).$ Let $f_n(x) = sin (nx).$ Then for any $varphi in C_c^infty(0,2pi),$ $int_0^2pi f_n varphi to 0$ by Riemann-Lebesgue. Thus $f_nto 0$ in $D'(0,2pi),$ but certainly not in $L^2(02pi).$






        share|cite|improve this answer






















          up vote
          0
          down vote










          up vote
          0
          down vote









          Suppose $Omega = (0,2pi).$ Let $f_n(x) = sin (nx).$ Then for any $varphi in C_c^infty(0,2pi),$ $int_0^2pi f_n varphi to 0$ by Riemann-Lebesgue. Thus $f_nto 0$ in $D'(0,2pi),$ but certainly not in $L^2(02pi).$






          share|cite|improve this answer












          Suppose $Omega = (0,2pi).$ Let $f_n(x) = sin (nx).$ Then for any $varphi in C_c^infty(0,2pi),$ $int_0^2pi f_n varphi to 0$ by Riemann-Lebesgue. Thus $f_nto 0$ in $D'(0,2pi),$ but certainly not in $L^2(02pi).$







          share|cite|improve this answer












          share|cite|improve this answer



          share|cite|improve this answer










          answered Aug 30 at 22:18









          zhw.

          67.2k42872




          67.2k42872




















              up vote
              0
              down vote













              Another way to see this is the following:



              It is known that $D(Omega)$ is dense in $D'(Omega)$ and $L^2(Omega)$ with respect to their natural topologies. By considering delta functions, evidently $L^2(Omega) neq D'(Omega)$ if $Omega subset mathbb R^n$ is open.



              So if we take $u in D'(Omega) setminus L^2(Omega)$ and choose $u_n in D(Omega)$ such that $u_n rightarrow u$ in $D'(Omega),$ I claim $(u_n)$ does not converge in $L^2(Omega).$ Indeed if $u_n rightarrow v$ in $L^2(Omega),$ it converges weakly in particular and hence $u_n rightarrow v in D'(Omega).$ So $u = v in L^2(Omega),$ which is a contradiction.






              share|cite|improve this answer
























                up vote
                0
                down vote













                Another way to see this is the following:



                It is known that $D(Omega)$ is dense in $D'(Omega)$ and $L^2(Omega)$ with respect to their natural topologies. By considering delta functions, evidently $L^2(Omega) neq D'(Omega)$ if $Omega subset mathbb R^n$ is open.



                So if we take $u in D'(Omega) setminus L^2(Omega)$ and choose $u_n in D(Omega)$ such that $u_n rightarrow u$ in $D'(Omega),$ I claim $(u_n)$ does not converge in $L^2(Omega).$ Indeed if $u_n rightarrow v$ in $L^2(Omega),$ it converges weakly in particular and hence $u_n rightarrow v in D'(Omega).$ So $u = v in L^2(Omega),$ which is a contradiction.






                share|cite|improve this answer






















                  up vote
                  0
                  down vote










                  up vote
                  0
                  down vote









                  Another way to see this is the following:



                  It is known that $D(Omega)$ is dense in $D'(Omega)$ and $L^2(Omega)$ with respect to their natural topologies. By considering delta functions, evidently $L^2(Omega) neq D'(Omega)$ if $Omega subset mathbb R^n$ is open.



                  So if we take $u in D'(Omega) setminus L^2(Omega)$ and choose $u_n in D(Omega)$ such that $u_n rightarrow u$ in $D'(Omega),$ I claim $(u_n)$ does not converge in $L^2(Omega).$ Indeed if $u_n rightarrow v$ in $L^2(Omega),$ it converges weakly in particular and hence $u_n rightarrow v in D'(Omega).$ So $u = v in L^2(Omega),$ which is a contradiction.






                  share|cite|improve this answer












                  Another way to see this is the following:



                  It is known that $D(Omega)$ is dense in $D'(Omega)$ and $L^2(Omega)$ with respect to their natural topologies. By considering delta functions, evidently $L^2(Omega) neq D'(Omega)$ if $Omega subset mathbb R^n$ is open.



                  So if we take $u in D'(Omega) setminus L^2(Omega)$ and choose $u_n in D(Omega)$ such that $u_n rightarrow u$ in $D'(Omega),$ I claim $(u_n)$ does not converge in $L^2(Omega).$ Indeed if $u_n rightarrow v$ in $L^2(Omega),$ it converges weakly in particular and hence $u_n rightarrow v in D'(Omega).$ So $u = v in L^2(Omega),$ which is a contradiction.







                  share|cite|improve this answer












                  share|cite|improve this answer



                  share|cite|improve this answer










                  answered Aug 31 at 2:20









                  ktoi

                  1,7061615




                  1,7061615



























                       

                      draft saved


                      draft discarded















































                       


                      draft saved


                      draft discarded














                      StackExchange.ready(
                      function ()
                      StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f2899105%2fdoes-convergence-in-d-omega-imply-convergence-in-l2-omega%23new-answer', 'question_page');

                      );

                      Post as a guest













































































                      這個網誌中的熱門文章

                      How to combine Bézier curves to a surface?

                      Mutual Information Always Non-negative

                      Why am i infinitely getting the same tweet with the Twitter Search API?