In college, I've come across many instances where we multiply a derivative by a function, and the result somehow becomes the derivative of the function i.e $\frac{d}{dx}\times f=\frac{df}{dx}$— as if we're multiplying "operators" with functions in a purely algebraic way. This has always puzzled me, because it seems to work, but I don't fully understand why it works.
Here are a few examples to illustrate my confusion:
1. Curl of a vector field:
$$ \text{curl } \mathbf{F} = \nabla \times (M, N, P) = \left( \frac{\partial}{\partial x}, \frac{\partial}{\partial y}, \frac{\partial}{\partial z} \right) \times (M, N, P) $$
When evaluating the determinant in the cross product definition, it seems like we’re "multiplying" the components of a vector of operators (whatever that means) with the components of a vector field. But that feels mysterious. Why is it okay to do something like:
$$ \left( \frac{\partial}{\partial x} \right) \cdot M $$
and call the result just $\frac{\partial M}{\partial x}$? What allows us to manipulate the gradient operator $\nabla$ as though it were a vector?
If this is valid, why can't I do something like:
$$\operatorname{ pie} (M,N,P) = (!, \sqrt{\ }, \wedge 2) \times (M, N, P) $$
where “!” is the factorial operator, and $\wedge 2$ is the square operator, just for the sake of argument? it doesn't make less or more sense that the defintion of curl in my opinion.
2. Divergence of a vector field:
$$ \text{div } \mathbf{F} = \nabla \cdot (M, N, P) = \left( \frac{\partial}{\partial x}, \frac{\partial}{\partial y}, \frac{\partial}{\partial z} \right) \cdot (M, N, P) $$
Same idea — we’re dotting a vector of operators with a vector of functions. What justifies treating the differential operator as if it were a vector component that can be dotted and crossed in this way?
3. Operator methods in ODEs:
In solving linear differential equations like:
$$ a y'' + b y' + c y = f $$
we often introduce the differential operator $ D = \frac{d}{dx} $ and rewrite the equation as:
$$ (aD^2 + bD + c)y = f $$
Then we manipulate this expression algebraically, even dividing by the operator polynomial, e.g.,
$$ y = \frac{1}{aD^2 + bD + c} f $$
In some cases, we go even further and treat operator expressions like $\frac{1}{1 - D}$ (I also have no idea how to define $1/(1-D)$) as a geometric series:
$$ \frac{1}{1 - D} = 1 + D + D^2 + \dots $$
and apply this to a function $f$, obtaining:
$$ f + Df + D^2f + \dots $$
How does this make sense? Why are we allowed to expand differential operators like power series and apply them this way?
More about my doubts about this method here.
4. Schrödinger equation and time evolution:
In quantum mechanics, I remember (though somewhat vaguely) that we would manipulate expressions like:
$$ i\hbar \frac{\partial}{\partial t} \psi = H \psi $$
and treat $\frac{\partial}{\partial t}$ almost as an algebraic quantity. At one point, I recall solving for $4\psi(t)$ by exponentiating the Hamiltonian operator and applying it to $\psi(0)$, like:
$$ \psi(t) = e^{-iHt/\hbar} \psi(0) $$
This again treats differential operators as algebraic objects, able to be exponentiated, composed, and manipulated.
We also did this :- $\frac{ \partial w(x,t)}{ \partial x} =\frac{ \partial }{ \partial x}\times \cancel{ w(x,t)} =\frac{2i\pi \cancel{ w(x,t)} P}{h} $ To get $\frac{ \partial}{ \partial x}= \frac{2i\pi P}{h} $
My Question:
All of these cases seem to rely on treating differential operators like algebraic objects — manipulating them in ways that resemble ordinary algebra (adding, multiplying, factoring, even inverting or expanding them in power series). Why is this allowed? What mathematical framework makes this rigorous?
I've asked professors and classmates, but usually get vague or unsatisfying answers like "it just works" or "it's a notation thing.". I feel there must be a more rigorous explanation but I don't know where to look.
Is there a general theory or justification that explains why and when it’s valid to treat differential operators this way?