20

We let $\pi$ be the bijection between coefficients of the real monic polynomials to the real monic polynomials. Let $a\in \mathbb R^n$ be fixed vector. Then \begin{align*} \pi(a) = t^n + a_{n-1} t^{n-1} + \dots + a_0. \\ \end{align*} Now denote the set $$ \Delta = \{ x\in \mathbb R^n: \pi(x) \text{ has roots in the open unit disk of } \mathbb C\}.$$ It can be shown $\Delta$ is a path-connected set by Vieta's formula (maybe slightly modified for the real case). Let us consider a line (1-dim subspace) in $\mathbb R^n$, $L = \alpha b$ where $\alpha \in \mathbb R$ and $b \in \mathbb R^n$ is fixed. Clearly $L \cap \Delta$ is nonempty since $0 \in L \cap \Delta$.

I am trying to determine the number of connected components of $L \cap \Delta$.

We shall assume $n > 2$. If $n=1$, $L \cap \Delta$ is clearly connected. For $n=2$, if I am not mistaken, the paper by Fell https://projecteuclid.org/euclid.pjm/1102779366#ui-tabs-1 asserts that convex combination of real monic polynomials of the same degree with roots in the unit disk remains in the unit disk (Theorem 4). This allows us to construct a path between any $x \in L \cap \Delta$ and $0 \in \mathbb R^n$ and so $L \cap \Delta$ is connected.


Edit: The question was initially unclear in the sense: I didn't know whether $L \cap \Delta$ is connected. @Jean-Claude Arbaut gave nice numerical examples and plot to show that the set is not connected. I rewarded a bounty and started a new one to see whether there is some bound of the number of connected components. I would reward the bounty to any bound or if the number of connected components could be unbounded.

user1101010
  • 3,638
  • 1
  • 17
  • 40
  • I don't understand your last assertion. Take $P_0=x^2-1,P_1=x+1$. Consider the polynomial $(1-u)P_0+uP_1$ when $u$ tends to $1$. –  Nov 20 '18 at 19:57
  • I am considering monic polynomials of the same degree. – user1101010 Nov 20 '18 at 19:58
  • Can you reproduce the argument in your linked article ? Also let $f(r,z) = z^n+r \sum_{k=0}^{n-1}a_k z^k$ and $(t,Z(t))$ be the curve such that $f(t,Z(t)) = 0$ with $Z(t)$ the largest root of $f(t,z)$. Set $h(t) = |Z(t)|^2$. You want to look at the zeros of $h(t) - 1$ and $h'(t)$. – reuns Nov 23 '18 at 07:18
  • You are telling you are trying to prove something and did not find a counterexample, but it's not clear what you are trying to prove. Anyway here is a numerical example showing that there can be more than one component: $(x-0.9)^3$. When you multiply the corresponding $a$ by a value in the range $[0.708,0.983]$, you get outside of the unit circle. – Jean-Claude Arbaut Nov 23 '18 at 10:12
  • @reuns: Sorry I completely missed your comment. Should we consider $Z(t)$ as the largest root modulus of $f(t, z)$? In this case, is $h(t) = |Z(t)|^2$ still a polynomial in $t$? Thanks. – user1101010 Nov 26 '18 at 05:47
  • @reuns: I realized $h(t)$ is not a polynomial in $t$. It seems to me the behavior of this function could be quite arbitrary. Could you provide further insights on your comment? Thanks. – user1101010 Nov 26 '18 at 18:33

2 Answers2

11

Here is an illustration of the comment above, and another example.

The set $L\cap\Delta$ considered here is a subset of a line, or if we consider a parametrization, a subset of $\Bbb R$. The connected components are thus intervals.

I think the main question is: how many intervals are there?


First, a plot of the roots of the polynomial $(x-0.9)^3$ and of the subsequent polynomials when you multiply $a_0\dots a_2$ by $\lambda\in[0,1]$. The roots follow three differents paths, and eventually get closer to zero, but two of them first get out of the unit circle. Note that here I consider only $\lambda\in[0,1]$ and not $\lambda\in\Bbb R$, but for all the examples below, that does not change the number of connected components.

enter image description here


Another example, of degree $9$, with initially three roots with multiplicity $3$ each. The roots are $0.95$, $0.7\exp(2i\pi/5)$ and $0.7\exp(-2i\pi/5)$. The colors show different portions of the paths, so that you can see there will be $3$ connected components (which correspond to the black portions below).

enter image description here


Here is an example with $4$ components.

The initial roots are $0.95$, $0.775\exp(\pm0.8482i)$, $0.969\exp(\pm2.7646i)$ with multiplicity $3,2,2$ respectively. Note the behaviour of "root paths" is highly sensitive to the initial roots.

enter image description here

How to read this ($\lambda$ decreases from $1$ to $0$ in the successive steps):

  • Initially ($\lambda=1$), all roots are inside the unit circle, and we are on a connected component of $L\cap\Delta$.
  • The first roots to get out are in red. The other ones are still inside the unit circle.
  • The "red roots" get inside the unit circle: second component.
  • Now some of the "blue roots" get out.
  • The "blue roots" get in, and all roots are inside the unit circle: third component.
  • Some "green roots" get out.
  • The "green roots" get in, and all roots are inside the unit circle: fourth and last component, and after that the roots converge to zero as $\lambda\to0$.

Now, could there be more components? I have not a proof, but my guess would be that by cleverly choosing the initial roots, it's possible to get paths that will get outside then inside the unit circle in successive order, and the number of components could be arbitrary. Still investigating...


R program to reproduce the plots (as is, the last plot).

# Compute roots given vector a in R^n and coefficient e
# That is, roots of $x^5 + e a_n x^{n-1} + \cdots + e a_0$
f <- function(a, e) {
  polyroot(c(a * e, 1))
}

# Given vector a and number of points, compute the roots for
# each coefficient e = i/n for i = 0..n.
# Each set of root get a color according to:
# * if |z|<1 for all roots, then black
# * otherwise reuse the preceding color (and change if
#   the preceding was black)
# Return in z the list of all roots of all polynomials,
# and in cl the corresponding colors.
mk <- function(a, n) {
  cls <- c("red", "blue", "green", "yellow")
  z <- NULL
  cl <- NULL
  cc <- "black"
  k <- length(a)
  j <- 0
  for (i in n:0) {
    zi <- f(a, i / n)
    if (all(abs(zi) <= 1)) {
      cc <- "black"
    } else {
      if (cc == "black") {
        j <- j + 1
        cc <- cls[j]
      }
    }
    z <- c(z, zi)
    cl <- c(cl, rep(cc, k))
  }
  list(z=z, cl=cl)
}

# Compute polynomial coefficients from roots
pol <- function(a) {
  p <- c(1)
  for (x in a) {
    p <- c(0, p) - c(x * p, 0)
  }
  p
}

# New plot, and draw a circle
frame()
plot.window(xlim=c(-1.0, 1.0), ylim=c(-1.0, 1.0), asp=1)
z <- exp(2i  *pi * (0:200) / 200)
lines(z, type="l", col="darkgreen", lwd=2)

# The third example given
a <- c(0.95, 0.775 * exp(0.8482i), 0.775 * exp(-0.8482i),
       0.969 * exp(2.7646i), 0.969 * exp(-2.7646i))
# Duplicate roots, compute coefficients, remove leading x^n
a <- head(pol(rep(a, times=c(3, 2, 2, 2, 2))), -1)

# Plot roots
L <- mk(a, 3000)
points(L$z, col=L$cl)
Jean-Claude Arbaut
  • 23,601
  • 7
  • 53
  • 88
  • Thanks for your nice demonstrations. Yes, my initial question was unclear since I didn't know whether they are connected or not (too blind to see your examples.) – user1101010 Nov 24 '18 at 00:26
  • May I ask how you come up with these numerical examples in the first place? Is there intuition behind these examples? – user1101010 Nov 24 '18 at 00:27
  • Your demonstrations are very nice and clear. Probably you don't care, but I will reward the bounty to you when the system allows me to and start a new one to see whether there is a systematic way to investigate the number of connected components. Thanks. – user1101010 Nov 24 '18 at 00:32
  • @user9527 My first thought was that maybe there would be only one component. Luckily the first example I tried ($(x-0.9)^3$), with a multuiple root, made things much clearer: as the three (initially common) roots depart, they go in different directions and thus can get out of the circle. Then it's just a matter of playing with the location of the roots. – Jean-Claude Arbaut Nov 24 '18 at 09:21
  • @user9527 There are still many unknowns. The number of components is certainly bounded for a fixed $n$, and it would be interesting to find a tight bound. Also, it would be interesting to know more about the geometry of locii of lines for which the number of components is a given $k$ (it was hard to find an example "by hand" for $k=4$ components, which makes me think for larger $k$ the polynomials become more and more rare). And I have no proof the number of components is unbounded for varying $k$. There could be a much better/simpler approach too. – Jean-Claude Arbaut Nov 24 '18 at 10:10
  • Thanks. How do we argue the connected components are finite for fixed $n$? – user1101010 Nov 25 '18 at 23:34
  • Hi Jean-Claude. I noticed in your plots, the $0$ roots seemed to be isolated. Should the roots be a continuous function over the parametrization of the line? – user1101010 Dec 01 '18 at 06:39
  • @user9527 It's due to the discrete range of parameters I took: it was $\lambda\in{k/n, k=0\dots n}$, and the roots seem to move much "faster" when near $0$ (and also when near another multiple root), hence the apparent "jumps". By the way, if you want to play with these plots, I can give you the R programs (it can be done with many other languages, of course). – Jean-Claude Arbaut Dec 01 '18 at 09:40
  • Thanks. That would be great. I have tried to produce your plots in python. But I do want to see the tricks you use to color those paths. – user1101010 Dec 01 '18 at 18:07
  • @user9527 Here you go :) – Jean-Claude Arbaut Dec 01 '18 at 19:20
  • Thank you so much. – user1101010 Dec 01 '18 at 23:44
3

I think Cauchy's argument principle can provide an upper bound(possibly not tight). As is well-known, Cauchy's argument principle says that for a polynomial $p(z)$ which does not vanish on $|z|=1$, its number of roots counted with multiplicity in $|z|<1$ is given by the formula $$ \frac{1}{2\pi i}\int_{|z|=1} \frac{p'(z)}{p(z)}dz. $$ Let us denote $p_\alpha (z) = \pi(\alpha b)(z)$ for $\alpha \in\mathbb{R}$ and fixed $b\in\mathbb{R}^n$. We will use the integer-valued continuous function $$ F:\alpha \mapsto \frac{1}{2\pi i}\int_{|z|=1} \frac{p_\alpha'(z)}{p_\alpha(z)}dz, $$ defined for $\alpha$ such that $p_\alpha$ does not vanish on the unit circle. However, we can easily see that the number of such values of $\alpha$ is at most $n+1$.
Assume $p_\alpha(e^{-i\theta})= 0$ for some $\theta$. Then, $$ 1+ \alpha(b_{n-1}e^{i\theta} +b_{n-2}e^{i2\theta} +\cdots +b_0e^{in\theta}) = 0\quad\cdots(*). $$Then, in particular, $b_{n-1}e^{i\theta} +b_{n-2}e^{i2\theta} +\cdots +b_0e^{in\theta}$ is real-valued, and hence its imaginary part $$ \sum_{k=1}^n b_{n-k}\sin(k\theta) = 0. $$ If we denote $T_n$ by $n$-th Chebyshev polynomial, then it holds $\sin\theta\cdot T_n'(\cos\theta) = n\sin n\theta$. Here, $U_{n-1}:= \frac{1}{n}T_n'$ is a polynomial of degree $n-1$. From this, we may write $$ \sin\theta\sum_{k=1}^n b_{n-k}U_{k-1}(\cos\theta) =\sin\theta \cdot P(\cos\theta)= 0, $$ where $P = \sum_{k=1}^n b_{n-k}U_{k-1}$ is a polynomial of degree at most $n-1$. Thus, solving the above equation, we get $\theta = 0\text{ or }\pi$, or $\cos\theta= v_1, v_2,\ldots v_{n-1}$, where each $v_i$ is a root of $P=0$.
Now, from $(*)$, we have $$ 1+ \alpha(b_{n-1}\cos\theta +b_{n-2}\cos2\theta +\cdots +b_0\cos n\theta) = 0, $$ and the number of possible values of $\alpha$ is at most $n+1$. Let these values $-\infty =:\alpha_0 <\alpha_1<\alpha_2< \ldots< \alpha_N<\alpha_{N+1}:=\infty$. Then, $F$ is well-defined on each $(\alpha_j, \alpha_{j+1})$ and is a continuous, integer-valued function. Thus, $F \equiv m_j$ on each $(\alpha_j, \alpha_{j+1})$. Our next goal is to investigate how the component of $L\cap \Delta$ looks like using the information of $F$.

(i) Notice that $\alpha_j \neq 0$ for all $j$ and there is unique $j'$ such that $0\in (\alpha_{j'}, \alpha_{j'+1})$. Clearly, on $(\alpha_{j'}, \alpha_{j'+1})$, $F$ is identically $n$.

(ii) On the other hand, note that $\lim_{|\alpha|\to\infty} F(\alpha) \leq n-1$ since as $|\alpha| \to \infty$, the roots of $p_\alpha$ acts like that of $n-1$-degree polynomial.

(iii) If $F = n$ on adjacent intervals $(\alpha_j, \alpha_{j+1})$ and $(\alpha_{j+1}, \alpha_{j+2})$, then we can conclude that on $\alpha = \alpha_{j+1}$, $p_\alpha$ has all its roots on the "closed" unit disk, since on a punctured neighborhood of $\alpha_{j+1}$, all the roots are contained in the open unit disk, hence in the closed one.(Note that the behavior of zero set is continuous.) Thus we conclude in this case, $(\alpha_j, \alpha_{j+1})$ and $(\alpha_{j+1}, \alpha_{j+2})$ are contained in the same component (together with $\alpha_{j+1}$).

(iv)There is possibility that all the roots of $p_{\alpha_{j+1}}$ are contained in the closed unit disk (and of course some of them are on the boundary), even though $F<n$ on $(\alpha_j, \alpha_{j+1})$ and $(\alpha_{j+1}, \alpha_{j+2}).$ In this case the singleton set $\{\alpha_{j+1}\}$ forms a component.

Summing this up, we can see that the components of $L\cap \Delta$ is of the form $(\alpha_i, \alpha_{j})$, $[\alpha_i, \alpha_{j})$, $(\alpha_i, \alpha_{j}]$, or $[\alpha_i, \alpha_{j}]$ for some $i\leq j$ (if equal, then it becomes sigleton set). And by (iii), $(\alpha_i, \alpha_{j})$ and $(\alpha_j, \alpha_{k})$ cannot be separated, and one of its component should include non-singleton component containing $0$. From this, the number of components can be greatest when all the components are of the form $\{\alpha_j\}$ except for one $(\alpha_{j'},\alpha_{j'+1})$. Hence we conclude that the number $k$ of components satisfy $$k \leq N-1 \leq n.$$

$\textbf{EDIT:}$ The above argument is valid if it were that $$\Delta = \{ x\in \mathbb R^n: \pi(x) \text{ has roots in the }\textbf{closed} \text{ unit disk of } \mathbb C\}.$$ There was a slight revision after I wrote. However, the argument can be easily modified to give the same bound $k\leq n$ for the case $$\Delta = \{ x\in \mathbb R^n: \pi(x) \text{ has roots in the open unit disk of } \mathbb C\}. $$

$\textbf{EDIT:}$ I've thought about when it happens that some paths of roots touch the boundary and get back to the interior. As a result, I could reduce the previous bound $n$ to $\frac{n+1}{2}$.
Assume that on $F=n$ on $(\alpha_{j-1}, \alpha_{j})\cup (\alpha_{j}, \alpha_{j+1}).$ By definition, $p_{\alpha_j}(z^*)=0$ for some $|z^*|=1$. The first claim is that the multiplicity of $z^*$ is $1$. Here is heuristic argument. Assume $z^*$ has multiplicity $L$. Then, we have for $q(z) = b_{n-1}z^{n-1} + b_{n-2}z^{n-2} + \cdot + b_0$, $$ p_\alpha(z) = p_{\alpha_j}(z) + (\alpha-\alpha_j)q(z) = (z-z^*)^L r(z) + (\alpha-\alpha_j)q(z), $$ where $r(z)$ is a polynomial s.t. $r(z^*) \neq 0$. As $\alpha \sim \alpha_j$, we have $z \sim z^*$, and solving $p_\alpha = 0$ is asymptotically equivalent to $$ (z-z^*)^L \sim - (\alpha-\alpha_j)\frac{q(z^*)}{r(z^*)}. $$Note that $q(z^*) \neq 0$ since if $q(z^*) = 0$, then $p_{\alpha_j}(z^*) = 0$ implies $z^* = 0$, leading to contradiction. Let $\zeta_L$ be $L$-th root of unity and $\omega^{\frac{1}{L}}$ denote one of the ($L$)-solutions of $z^L= \omega$. This shows $$ \lambda_k(\alpha) \sim z^* + \left(- (\alpha-\alpha_j)\frac{q(z^*)}{r(z^*)}\right)^{\frac{1}{L}}\zeta^k_L, \quad k=1,2,\ldots, L, $$are asymptotic roots of $p_\alpha (z) = 0$. We will see that as $\alpha \uparrow \alpha_j$ or $\alpha \downarrow \alpha_j$, it is impossible that all the $\lambda_k(\alpha)$ lie in the open unit disk.
Formal proof of this claim asserting that an analytic function that has $L$-th zero behaves locally like $L$-to-$1$ function requires a version of Rouche's theorem and argument principle. Assume an analytic function $f$ has zero of $L$-th order at $z=0$. For sufficiently small $\epsilon$, $f$ does not vanish on $0<|z| \leq \epsilon$, and $$ \frac{1}{2\pi i}\int_{|z|=\epsilon} \frac{f'(z)}{f(z)}dz = L $$ gives the number of zeros on $|z|<\epsilon$. If we perturb $f$ by $\eta \cdot g(z)$ as $\eta \to 0$, then $$ \frac{1}{2\pi i}\int_{|z|=\epsilon} \frac{f'(z)+\eta g'(z)}{f(z)+\eta g(z)}dz = L $$ for $|\eta| < \frac{\min_{|z|=\epsilon}|f(z)|}{\max_{|z|=\epsilon}|g(z)|}$.(This is by Rouche's theorem.) Thus, small perturbation does not affect the number of zeros on a neighborhood of $0$.

Let us consider the equation $$ z^Lf(z) = u^L\quad\cdots (***), $$where $f$ is analytic, $f(0) = 1$ and $u\in\mathbb{C}$ is an $\mathcal{o}(1)$ quantity (this means $|u|\to 0$.) If $f = 1$, then the exact roots of the equation is $$ \lambda_k = u\zeta^k,\quad k=1,2,\ldots, L, $$where $\zeta$ is the $L$-th root of unity. Our claim about asymptotic roots of $(***)$ is that $$ \lambda_k(u) = u(\zeta^k + \mathcal{o}_u(1)),\quad k=1,2,\ldots, L, $$ is the roots of $(***)$. Here, $\mathcal{o}_u(1)$ denotes some quantity going to $0$ as $|u|\to 0$. Proof is simple. Modify the equation $(***)$ to $$ z^L f(uz) = 1. $$ Then $u\lambda'_k, k=1,2,\ldots, L$ is the roots of $(***)$ where $\lambda'_k$ denotes roots of modified equation. But as $|u|\to 0$, the modified equation converges to $z^L = 1$ whose exact roots are $\zeta^k, k=1,2,\ldots, L.$ By Rouche's theorem, each $\lambda'_k$ should be located in a neighborhood of $\zeta^k$. This proves the claim.
Actually this implies a seemingly stronger assertion that if $$ z^Lf(z) = v^L $$ where $v = u(1+ \mathcal{0}_u(1))$, then $$ \lambda_k = u(\zeta^k + \mathcal{o}_u(1)). $$ And we will use this version. We are assuming that $$ (z-z^*)^L \frac{r(z)}{r(z^*)} = -(\alpha-\alpha_j)\frac{q(z)}{r(z^*)}. $$ As $\alpha \to \alpha_j$, the roots $z\to z^*$ by Rouche's theorem. Thus, $$ -(\alpha-\alpha_j)\frac{q(z)}{r(z^*)} = -(\alpha-\alpha_j)\left(\frac{q(z^*)}{r(z^*)}+\mathcal{o}_{\alpha-\alpha_j}(1)\right). $$ By the above claim, we get $$ \lambda_k(\alpha) = z^* + \left(- (\alpha-\alpha_j)\frac{q(z^*)}{r(z^*)}\right)^{\frac{1}{L}}(\zeta^k + \mathcal{o}_{\alpha-\alpha_j}(1)) $$ as claimed.

We can see that if $L\geq 2$, then $\lambda_k, k=1,2,\ldots, L$ comes from $L$ different directions and they are equally spaced. If $L\geq 3$, we can easily see that it is impossible for all the $\lambda_k$ to lie in the unit circle. The case $L=2$ is more subtle, but we can see that it is also impossible in this case. (By noting that the unit circle has a positive curvature at each point.) Hence, $L$ should be $1$. We now know that $z^*$ should have multiplicity $1$, and $$ \lambda(\alpha) = z^* -(\alpha-\alpha_j)\left(\frac{q(z^*)}{r(z^*)} +\mathcal{o}_{\alpha-\alpha_j}(1)\right). $$ We see that $r(z^*) = p'_{\alpha_j}(z^*)$ by definition, and $\frac{\partial}{\partial \alpha}\lambda(\alpha_j) = - \frac{q(z^*)}{p'_{\alpha_j}(z^*)}$. For $\lambda(\alpha)$ to get back to the inside, it must be that $$ \frac{q(z^*)}{p'_{\alpha_j}(z^*)} = -\frac{\partial}{\partial \alpha}\lambda(\alpha_j) = i\beta z^* $$ for some real $\beta$ (that is, tangent vector should be orthogonal to the normal vector $z^*$.) From now on, let us write $\alpha_j$ as $\alpha^*$ for notational convenience.
Hence, we have $q(z^*) = i\beta z^* p_{\alpha^*}'(z^*) $. From $p_{\alpha^*}(z^*) = 0$, we also have $$(z^*)^n = -\alpha^* q(z^*) = -i\alpha^*\beta z^*p'_{\alpha^*}(z^*)=-i\alpha^*\beta z^*(n(z^*)^{n-1} + \alpha^*q'(z^*)). $$ Hence, $(1+i\alpha^*\beta n)(z^*)^{n} = -i(\alpha^*)^2\beta z^*q'(z^*),$ and we get $$ (z^*)^n = \frac{-i(\alpha^*)^2\beta}{1+i\alpha^*\beta n}z^*q'(z^*) = -\alpha^*q(z^*), $$ $$ \frac{i\alpha^*\beta}{1+i\alpha^*\beta n}z^*q'(z^*) =q(z^*). $$ Note that it holds that $$ \frac{z^*q'(z^*)}{q(z^*)} = n -\frac{i}{\alpha^*\beta}. $$(Note that $q(z^*) \neq 0 $ and $\beta \neq 0$.) If we take conjugate on both sides, since $q(z)$ is a real polynomial, we have $$ \frac{\overline{z^*}q'(\overline{z^*})}{q(\overline{z^*})} = n +\frac{i}{\alpha^*\beta}. $$ This means $z^*$ cannot be $\pm 1$.
To derive an equation about $z^*$, let us write $z^* = e^{-i\theta}$. Then we have from $(z^*)^n= -\alpha^*q(z^*)$, that $$ 1 = -\alpha^*\left(\sum_{k=1}^{n} b_{n-k} e^{ik\theta}\right). $$Also from $(z^*)^n = \frac{-i(\alpha^*)^2\beta}{1+i\alpha^*\beta n}z^*q'(z^*)$, we have $$ \frac{1+i\alpha^*\beta n}{i\alpha^*\beta}=n - \frac{i}{\alpha^*\beta} = -\alpha^*\left(\sum_{k=1}^{n} (n-k)b_{n-k} e^{ik\theta}\right). $$ Both equations together yields: $$ 1 = -\alpha^*\left(\sum_{k=1}^{n} b_{n-k} e^{ik\theta}\right), $$ $$ \frac{i}{\alpha^*\beta} = -\alpha^*\left(\sum_{k=1}^{n} kb_{n-k} e^{ik\theta}\right). $$ Take the imaginary part for the former and the real part for the latter. Then, $$ \sum_{k=1}^{n} b_{n-k} \sin(k\theta)=\sin\theta\cdot P(\cos\theta) = 0, $$ $$ \sum_{k=1}^{n} kb_{n-k} \cos(k\theta)=\cos\theta\cdot P(\cos\theta) -\sin^2\theta\cdot P'(\cos\theta)=0, $$ where $P := \sum_{k=1}^n b_{n-k}U_{k-1}$ is a polynomial of degree at most $n-1$. We already know $\sin\theta \neq 0$ since $e^{i\theta}$ is not real. Thus, it must be that $P(\cos\theta) = P'(\cos\theta) = 0$, meaning that $P(v)$ has a multiple root at $v = \cos\theta\in (-1,1)$. Recall that the roots of $P(v)=0$ were used to prove that the number of possible $\alpha$'s is at most $n+1$. Each possible $\alpha$ was related to the root of the equation $\sin\theta \cdot P(\cos\theta) =0$ by $$ \alpha = -\frac{1}{b_{n-1}\cos\theta + \cdots + b_0\cos n\theta}. $$ Now, suppose $P(v)$ has roots $v_1,\ldots, v_k, w_1,\ldots, w_l$ in $(-1,1)$ where each $v_i$ has multiplicity $1$ and each $w_j$ is multiple roots. Then we must have $k + 2l \leq n-1.$ Some of $v_i$ and $w_j$ are mapped to $\alpha_i$ and $\beta_j$ via the above formula. Some may not because $b_{n-1}\cos\theta + \cdots + b_0\cos n\theta$ may vanish for $\cos\theta = v_i$ or $w_j$. Adding $-\frac{1}{b_{n-1} + \cdots + b_0}$ and $-\frac{1}{-b_{n-1} + \cdots + b_0(-1)^n}$ to the set of $\alpha_i$'s, we finally get $A = (\alpha_i)_{i\leq l'}, B=(\beta_j)_{j\leq k'}$ where $l' + 2k' \leq n+1$. We may assume that $A$ and $B$ are disjoint by discarding $\alpha_i$ such that $\alpha_i =\beta_j$ for some $j$. Order the set $A\cup B$ by $$\gamma_0=-\infty <\gamma_1 < \gamma_2 <\cdots< \gamma_{l'+k'} <\infty=\gamma_{l'+k'+1}.$$ If $\gamma_i\in A$, then on one of the $(\gamma_{i-1},\gamma_i)$ and $(\gamma_{i},\gamma_{i+1})$, $F$ should be $<n$. Let $R$ be the family of intervals $(\gamma_{i},\gamma_{i+1})$ on which $F<n$. It must include $(-\infty,\gamma_1)$ and $(\gamma_{l'+k'},\infty)$. And it follows that if $\gamma_i\in A$, then it must be one of the end points of some interval in $R$. This restricts the cardinality of $A$ by $$ l' = |A|\leq 2 + 2(|R|-2)=2|R|-2. $$ Note that $k'+l'+1-|R|$ is the number of intervals on which $F=n$. Thus, we have $$ k'+l'+1-|R|\leq k' + l' +1 -\frac{l'+2}{2}=k'+\frac{l'}{2}\leq \frac{n+1}{2}. $$, as we wanted.
To conclude, let us define the constant $C_n$ by $$ C_n = \max\{N\geq 1\;|\;\Delta\cap \{\alpha b\}_{\alpha\in\mathbb{R}}\text{ has }N\text{ components for some } b\in \mathbb{R}^n\}. $$ Then, we have $$ C_n \leq \frac{n+1}{2}.$$

(Note: Especially, for $n=2$, we have $C_2 = 1$ and hence $\Delta\cap\{\alpha b\}_{\alpha \in \mathbb{R}}$ is connected for every $b \in \mathbb{R}^2$.)

Myunghyun Song
  • 22,003
  • 2
  • 26
  • 62
  • May I ask what you mean by "closed" unit disk? Is it the boundary of the unit disk? Could you elaborate how you conclude that on $\alpha = \alpha_{j+1}$, $p_{\alpha}$ has all its roots on the "closed" unit disk? Thanks. – user1101010 Dec 03 '18 at 08:04
  • @user9527 The closed unit disk is just ${z\in \mathbb{C} ;|;|z|\leq 1}$. I stressed the term 'closed' because Cauchy's formula gives us the number of root inside the unit disk. To elaborate, note that the behavior of zero set of polynomial is continuous with respect to continuous perturbation. So, if zero set is contained in the closed unit disk on a neighborhood of $\alpha_{j+1}$, then it should also be true for $\alpha_{j+1}$. – Myunghyun Song Dec 03 '18 at 08:11
  • Thanks. I am not sure I understand correctly: are you showing that $F$ cannot be $n$ on two adjacent intervals? But why is this a contradiction? – user1101010 Dec 03 '18 at 08:26
  • @user9527 In fact, it can happen. What I showed is that if that is true, we can 'glue' two adjacent intervals to form a common component. – Myunghyun Song Dec 03 '18 at 08:42
  • Thanks. But I still don't see how we can 'glue' two intervals since if $p_{\alpha_{j+1}}$ has one root on the unit circle then $\alpha_{j+1} \notin L \cap \Delta$. – user1101010 Dec 03 '18 at 08:47
  • Maybe it was not clear enough in my question. In the title, I asked in the open unit disk. I edited the question to make it clearer. I meant to ask when $\Delta$ is an open unit disk. – user1101010 Dec 03 '18 at 08:50
  • Based on your answer, I guess the upper bound is still $n$, i.e., except the unbounded intervals, all other intervals are in the intersection $L \cap \Delta$ which is $N+1 - 2 = n$. However, it is hard to imagine this scenario would happen. – user1101010 Dec 03 '18 at 09:08
  • @user9527 I agree it's unlikely to happen even for $n=2$ case. It's still unsatisfactory bound. I'm seeing what happens in some concrete examples. Now I have a question. If you mean $\Delta$ by the open unit disk, then Theorem 4 in the paper by Fell (https://projecteuclid.org/euclid.pjm/1102779366#ui-tabs-1) seems somewhat irrelevant, doesn't it? – Myunghyun Song Dec 03 '18 at 09:31
  • That was a sloppy mistake I made in reading the paper and I overlooked the equal sign. Thanks for pointing out. But I think we can still work it through. – user1101010 Dec 03 '18 at 19:12
  • If we denote $\Gamma = {x \in \mathbb R^2: \pi(x) \text{ has roots in the closed unit disk}}$. $\Gamma$ is convex by Theorem $4$ of Fell's paper. Then the interior $\Gamma^{\circ}$ is convex. But we note for any point $x = (x_1, x_2) \in \Gamma \setminus \Delta$ with $\max( |r_1|, |r_2|)= 1$ where $r_1, r_2$ denote the roots and let $y(a) = \pi^{-1} ( (t-a r_1)(t-ar_2) )$ with $a \in [0, 1]$, clearly $y \in \Gamma$ if $a< 1$ and $\lim_{a \to 1}(y(a)) = x$. So this shows $\overline{\Delta} = \Gamma$. Since $\Delta$ is open, so $\Delta = \Gamma^{\circ}$ is convex. – user1101010 Dec 03 '18 at 19:12
  • @user9527 Oh, but $\overline{\Delta}=\Gamma$ implies $\Delta = \Gamma^o$ automatically? By the way, it seems that we need to look into how unit roots occurring at $\alpha = \alpha_j$ behave near the unit circle. If we can show that a root which touches the unit circle from inside cannot come back to inside, then the bound will be reduced to $\frac{n}{2}+\mathcal{O}(1)$ as I originally conjectured. – Myunghyun Song Dec 04 '18 at 00:55
  • You are right. It does not follow. Thanks. – user1101010 Dec 04 '18 at 04:24
  • Hi, Song. Thanks for your nice answer. In the sentence you newly edited, you said "By Rouche's theorem, each $\lambda'_k$ should be located in a neighborhood of $\zeta^k$.". Could you elaborate a little bit? I seem to miss something on this part. Thanks. – user1101010 Dec 07 '18 at 00:25
  • Sure. Consider $L$ disks centered at each $\zeta^k$ with radius $\epsilon>0$ sufficiently small. Then Rouche's theorem guarantees that as $|u|→0$, there is exactly one root $\lambda'_k$ in each disk. This means of course that $\lambda'_k = \zeta^k + \mathcal{o}_u(1)$. I hope this will help. – Myunghyun Song Dec 07 '18 at 00:50
  • Thanks for your time. Maybe I misunderstood but does your method shows the bound is $n$ even if the recurrent roots analysis does not work out? The condition should become $Im(q(\bar{z}) (z^n + a_{n-1} z^{n-1} + \dots + a_0) ) = 0$ but this would give a polynomial of degree $n$ with your clever Chebyshev trick. Or I missed something. – user1101010 Dec 14 '18 at 07:54
  • Oh, now I see it. What I did is as follows. We let $p(z) = z^n + a_{n-1}z^{n-1} + \cdots + a_0$ and suppose, for simplicity, that $p$ does not vanish on $|z|=1$. Then $p_\alpha(z) = p(z) +\alpha q(z)=0$ and also $p(\overline{z}) + \alpha q(\overline{z}) =0$. Since $\alpha$ cannot be $0$, and $\overline{z} = \frac{1}{z}$, we have $p( \frac{1}{z})q(z)-p(z)q( \frac{1}{z})=0$. This leads to a trigonometric polynomial of degree $\leq n$ and can give the same bound $n$ ! Anyway, good luck with your work ! – Myunghyun Song Dec 14 '18 at 08:15
  • Thanks for your answer. It’s very illuminating. – user1101010 Dec 14 '18 at 08:25
  • Hello Song. Sorry to bother you again. In another reading of proof, I think I didnot understand this part completely: If $L\geq 3$, we can easily see that it is impossible for all the $\lambda_k$ to lie in the unit circle. The case $L=2$ is more subtle, but we can see that it is also impossible in this case. (By noting that the unit circle has a tangent parabola at each point.) In particular, you said $L \ge 3$ is obvious, but I cannot see easily. And for $L=2$, you mentioned a tangent parabola, how does this help? Thank you. – user1101010 Dec 17 '18 at 18:41
  • Let's see for example $L=3$ case. Imagine $z^$ on the unit circle and draw $L=3$ rays from $z^$. If every angle between two adjacent rays is $\frac{2\pi}{3}$, then at least one of them should be outside the circle. This applies for all $L\geq 3$. When $L=2$, assume for simplicity that $z^* = i=(0,1)$. If we draw a straight line through $z^*$, then half of it should be outside the circle unless the straightline is $l:t+i, t\in\mathbb{R}$. So, if all $2$ roots are coming from the inside, their paths should be tangent to $l$. – Myunghyun Song Dec 17 '18 at 19:00
  • In view of the approximate root formula, it must be $\lambda(\alpha) \sim \pm \sqrt{|\alpha-\alpha_j|}+(i+\mathcal{o}(\sqrt{|\alpha-\alpha_j|}))$. But then its distance from 0 is $|\alpha-\alpha_j| + 1 +\mathcal{o}(|\alpha-\alpha_j|) >1$ for all sufficiently small $|\alpha-\alpha_j|$. It means both roots are outside the disk, and hence shows us that it is impossible that all roots lie in the disk as $\alpha\to \alpha_j$. Does this make it clear? I hope this will help. – Myunghyun Song Dec 17 '18 at 19:08
  • I see. This is very clear now. Thanks so much. – user1101010 Dec 17 '18 at 19:16