17

How does distribution theory plays role in solving differential equations? This question might seem to be very general. I will try to explain, please bear with me.

I understand, distributions make it possible to differentiate functions whose derivatives do not exist in the classical sense and any locally integrable function has a distributional derivative. In terms of differential equations, if coefficients of a differential operator are piece-wise continuous then we make use of distributions (how and why it works?).

I am more interested in their relation with Green's function. Please help me understand, how can I use distribution theory for solving differential equations.

pluton
  • 1,243
schzan
  • 571
  • 3
  • 15

3 Answers3

12

I'll try to present the role that distributions play when we're looking for solutions of differential equations. For the moment, let's work in $\mathbb{R}^n$ and with smooth functions only. Suppose that we would like to solve the Poisson equation $\Delta u = f$, given the function $f \in C^{\infty}$. If such a function $u \in C^{\infty}$ exists, then for any $v \in C^{\infty}_0$, we have \begin{equation} \int_{\mathbb{R}^n} v \Delta u = \int_{\mathbb{R}^n} fv. \tag{1} \end{equation} Using integration by parts, \begin{equation} -\int_{\mathbb{R^n}} \nabla u \cdot \nabla v = \int_{\mathbb{R}^n} fv. \tag{2} \end{equation} Let us now reverse the logic: suppose we can find a function $u$ that satisfies $(2)$ for every $v \in C^{\infty}_0$, then does it necessarily satisfy $\Delta u =f$? Whether or not it does, we'll call such a function $u$ a $\textbf{weak solution}$ of the differential equation $\Delta u = f$. It is in general much easier to show that a particular differential equation has a weak solution than a solution in the usual sense.

The idea of distributions takes this idea one step further: we can view the map $v \mapsto -\int_{\mathbb{R}^n} \nabla u \cdot \nabla v$ as a linear functional $C^{\infty}_0 \to \mathbb{R}$. We'll define a $\textbf{distribution}$ as a linear functional $C^{\infty}_0 \to \mathbb{R}$; as before, we can ask if we have a distribution $\nu$ that agrees with the linear functional $v \mapsto \int_{\mathbb{R}^n} fv$, can we build from it a solution of $\Delta u =f$? (Such a distribution will often also be called a weak solution.) While it is not clear whether or not this will always occur, this is the setting in with the machinery of functional analysis is best suited to tackle the problem.

In many nice cases (such as the Poisson equation for suitable $f$), if we have a distrubution that solves our PDE in the weak sense, then it will turn out that this is in fact a function, and its regularity will depend on that of the given function $f$ (for example, if $f \in C^{\infty}$ and $\Delta u =f$, then $u \in C^{\infty}$ as well; this is an example of elliptic regularity.)

The takeaway from the above is the following: we would like to find $u$ satisfying $\Delta u = f$ in some function space, so enlarge our function space to the space of distributions. There, show we can find a solution, and show that this solution in fact was a member of the original function space to start with (of course, this last sentence will not always work out, but this is the philosophy and the hope!).

msteve
  • 6,238
5

The most basic application is the use of the fundamental solution (also known as the Green's function) to solve inhomogeneous linear problems. When $*$ is convolution and $\delta$ is the Dirac delta centered at zero, $\delta * f=f$ for a wide class of $f$. On the other hand, if $L$ is a linear differential operator, then $Lu * f=L(u*f)$. (Or at least, this is definitely true when $u,f$ are smooth.) This means that if you could find a solution to $Lu=\delta$, then you could convolve it with $f$ on both sides to get $L(u*f)=f$. So $u*f$ is the solution to $Lv=f$ if $u$ is the solution to $Lw=\delta$. This $u$ is called the fundamental solution or Green's function for the operator $L$.

Duhamel's principle lets us extend this to time-dependent problems, provided the spatial differential operator is constant (and again linear). Cf. http://en.wikipedia.org/wiki/Duhamel%27s_principle#General_considerations

Ian
  • 104,572
  • Thank you for your answer. Can you please tell me why it works. How does distribution theory plays role here? – schzan Mar 08 '15 at 20:25
  • 2
    @shzan Which step would you like explained better? $\delta * f = f$ is essentially immediate from the definition of $\delta$ and the definition of the convolution of a distribution with a function. The definition of the convolution of a distribution with a function really is just a definition, but can be motivated by pretending that it is actually an integration and changing variables appropriately. – Ian Mar 08 '15 at 20:52
  • 2
    @shzan (Cont.) The fact that $Lu * f = L(u * f)$ for a linear differential operator $L$ and smooth functions $u,f$ is a routine exercise in exchanging derivatives and integrals. It is in fact the basis of the usual "mollification" techniques, where we approximate a nonsmooth function by a smooth function. The tricky matter is that $Lu * f=L(u * f)$ when $u$ is actually a distribution (and where we extend the domain of $L$ to distributions in an appropriate weak sense). Is that the step you're having trouble with? – Ian Mar 08 '15 at 20:53
  • 3
    @shzan (Cont.) The way we make sense of $Luf=L(uf)$ is a sneaky trick. In a convolution of smooth functions, we can move the derivatives around however we like: $L(fg)=Lfg=fLg$ when $f$ and $g$ are smooth. So we define $L$ on distributions in such a way that, if $u$ is a distribution and $f$ is a smooth function, $Luf=u*Lf$. Now this makes perfect sense as a convolution of a distribution with a smooth function. – Ian Mar 08 '15 at 20:59
  • 1
    @Ian $u*Lf$ is not always defined if $u$ is a distribution: at least one of distributions $u$ and $Lf$ has to be compactly supported. – TZakrevskiy Mar 08 '15 at 21:47
  • @TZakrevskiy I'm aware; I was being sloppy about "smooth", I really meant "appropriately nice". – Ian Mar 08 '15 at 22:17
  • @Ian Thank you so much for your explanation. Your third comment makes things very clear. Hopefully, $Lf$ is compactly supported in my case. Thank you guys for your help. You guys rock!! – schzan Mar 09 '15 at 10:09
  • @Ian Does this works even when the coefficients in $L$ are variables. Can you please mention any good reference book also. – schzan Mar 09 '15 at 14:16
  • @shzan I am nearly certain of the following. Assume $f \in C^\infty_c$, $L$ is a linear operator mapping $C^\infty_c$ into itself, and $Lu=\delta$. (Note that I am explicitly assuming that $Lu=\delta$ has a solution, which may impose some restrictions on $L$.) then $Luf=\deltaf=f$ and $Luf=uLf=L(uf)$, i.e. $uf$ solves $Lv=f$. – Ian Mar 09 '15 at 15:41
  • @Ian Thank you very much :), any reference? – schzan Mar 09 '15 at 15:45
  • @shzan As for references I honestly haven't found a very good one. Evans, the "standard" graduate PDE text, seems to be extremely averse to actually mentioning distributions explicitly. My professor at the moment is working out of Hormander's book The analysis of linear PDE. I have not read any of it, but he has notes online based on it, which are located here: http://www.math.umd.edu/~matei/ – Ian Mar 09 '15 at 15:45
  • @Ian Thanks again.:) – schzan Mar 09 '15 at 15:48
3

Below is the minimal mathematics - or rather the minimal Linear Systems Theory - needed to understand the role of distributions and Green's functions with differential equations. As extracted from old notes about "Signals and Systems".
Signals are represented by $\,x(t)\,$ and $\,y(t)\,$ where $\,x\,$ is an input signal / excitation, $\,y\,$ is an output signal / response and $\,t\,$ is time. A linear system $S$ is represented by $y(t) = S x(t)$ ; it produces an output when given an input and linearity means that: $$ S \left[ \lambda\, a(t) + \mu\, b(t) \right] = \lambda \, S \, a(t) \, + \, \mu \, S \, b(t) $$ So $S$ is a linear one-dimensional operator. More about linear operators and especially operators with (ordinary) differential equations in:

A system $S$ is homogeneous in time - also called Time invariant or Shift invariant - iff for all input signals $x(t)$ and for all output signals $y(t)$ and all time-shifts $\tau$ : $$ S \, x(t-\tau) = y(t-\tau) $$ Properties of linear homegeneous systems are for example: $$ S x'(t) = y'(t) \quad \Longleftrightarrow \quad S \frac{d}{dt} x(t) = \frac{d}{dt} S x(t) \quad \Longleftrightarrow \quad \left[ S \, , \, i\, \hbar \frac{d}{dt} \right] = 0 $$ The response of the derivative of the input is the derivative of the output; time differentiation cummutes with the operator of the system; conservation of energy is guaranteed (QM).
Consider the following summation, for a broad class of functions $f(t)$ : $$ S \left[ \sum_i f(\tau_i)\, x(t-\tau_i)\, \Delta\tau_i \right] = \sum_i f(\tau_i) \, S x(t-\tau_i) \, \Delta\tau_i = \sum_i f(\tau_i)\, y(t-\tau_i)\, \Delta\tau_i $$ Taking the limit of this Riemann sum for $\Delta\tau_i \to 0$ yields: $$ S \left[ \int_{-\infty}^{+\infty} f(\tau)\, x(t-\tau) d\tau \right] = \int_{-\infty}^{+\infty} f(\tau)\, y(t-\tau) d\tau $$ Where the convolution integrals $f * x$ and $f * y$ are recognized.
Suppose that a linear and homogeneous system is excited with a Dirac-delta as its input. Then the corresponding response is called a delta response, written as $\,h(t)\,$ by definition. So we have: $$ S\, x(t) = y(t) \qquad ; \qquad S\, \delta(t) = h(t) $$ This one-dimensional function is equivalent to Green's function when generalized to more dimensions e.g. space-time. The fundamental property of Dirac-delta says that: $x(t) = x(t) * \delta(t)$ , called "diafragma" property in Dutch, but couldn't find a nice English equivalent. Hence: $$ y(t) = S\, x(t) = S \left\{ \, x(t) * \delta(t) \, \right\} = x(t) * h(t) $$ Thus the superposition integral of $S$ has been found: $$ y(t) = h(t) * x(t) = \int_{-\infty}^{+\infty} h(\tau)\, x(t-\tau) d\tau $$ Consequently: if we know the Delta-response then we know any response of the system.
The above explains in a nutshell some essentials, at hand of one-dimensional linear & homogenous systems in time. I hope it nevertheless serves a purpose and that the reader is capable of thinking how to generalize this material to more than one dimension.
Han de Bruijn
  • 17,475