Skip to content
BY 4.0 license Open Access Published by De Gruyter June 26, 2023

Linear extension operators for Sobolev spaces on radially symmetric binary trees

  • Charles Fefferman EMAIL logo and Bo’az Klartag

Abstract

Let 1 < p < and suppose that we are given a function f defined on the leaves of a weighted tree. We would like to extend f to a function F defined on the entire tree, so as to minimize the weighted W 1 , p -Sobolev norm of the extension. An easy situation is when p = 2 , where the harmonic extension operator provides such a function F . In this note, we record our analysis of the particular case of a radially symmetric binary tree, which is a complete, finite, binary tree with weights that depend only on the distance from the root. Neither the averaging operator nor the harmonic extension operator work here in general. Nevertheless, we prove the existence of a linear extension operator whose norm is bounded by a constant depending solely on p . This operator is a variant of the standard harmonic extension operator, and in fact, it is harmonic extension with respect to a certain Markov kernel determined by p and by the weights.

MSC 2010: 46E36

1 Introduction

Consider a full binary tree of height N , whose set of vertices is denoted by

V = k = 0 N { 0 , 1 } k ,

i.e., the vertices are strings of zeroes and ones of length at most N . For x { 0 , 1 } k and k , we write π ( x ) { 0 , 1 } for the prefix of x of length . Thus, for k 1 , the parent of a vertex x { 0 , 1 } k V is the vertex π k 1 ( x ) . The set { 0 , 1 } 0 is a singleton whose unique element is denoted by , the empty string, which is the root of the tree. The set of leaves of the tree is { 0 , 1 } N , and all other vertices in V are internal vertices. A vertex x { 0 , 1 } k is said to have depth

d ( x ) = k ,

thus leaves have depth N and the root has depth 0. The collections of bijections from V to V that preserve depth and parenthood relations form a group. This group is referred to as the symmetry group of the tree. It has 2 2 N 1 elements, and it is a 2-Sylow subgroup of the group of all permutations of the leaves. The set

E = k = 1 N { 0 , 1 } k = V \ { }

is referred to as the set of edges of the tree, and the depth of an edge e { 0 , 1 } k is d ( e ) = k . That is, we think of e E as an undirected edge connecting the vertex whose string corresponds to e to its unique parent. Each internal vertex other than the root is connected to three vertices, which are its parent and its two children. Assume that we are given edge weights

W 1 , W 1 , , W N > 0 ,

where we view W k as the weight of all edges of depth k . For 1 < p < , the associated W ˙ 1 , p -seminorm is defined, for F : V R , via

(1) F W ˙ 1 , p ( V ) = k = 1 N W k x { 0 , 1 } k F ( x ) F ( π k 1 ( x ) ) p 1 / p .

We write V = { 0 , 1 } N V for the set of leaves of the tree. The trace of the W ˙ 1 , p -seminorm is defined, for f : V R , via

(2) f W ˙ 1 , p ( V ) = inf { F W ˙ 1 , p ( V ) ; F V = f } ,

i.e., the infimum of the W ˙ 1 , p -seminorm over all extensions of f from the leaves to the entire tree. We write R V for the collection of all functions f : V R , and similarly R V is the collection of all functions f : V R . Our main result is the following:

Theorem 1.1

Let 1 < p < and let W 1 , , W N > 0 . Then there exists a linear operator H : R V R V with the following properties:

  1. It is a linear extension operator, i.e., ( H f ) ( x ) = f ( x ) for any x V and any function f : V R .

  2. Its norm is bounded by a constant C ¯ p depending only on p , i.e., for any f : V R ,

    H f W ˙ 1 , p ( V ) C ¯ p f W ˙ 1 , p ( V ) .

In fact, we have the bound

(3) C ¯ p 4 p 1 / p q 1 / q ( 1 + max { ( p 1 ) 1 / p , ( q 1 ) 1 / q } ) C max 1 p 1 , 1 q 1 ,

where q = p / ( p 1 ) and where C > 0 is a universal constant.

The proof of Theorem 1.2 is constructive, and the extension operator H that we construct is in fact a harmonic extension operator with respect to a certain random walk defined on the tree. At each step, the random walk jumps from a vertex to one of its neighbors, where of course the neighbors of a vertex are its parent and its children. The Markov kernel corresponding to the random walk is invariant under the symmetries of the tree; thus, the probability to move from a vertex to its neighbor depends only on the weights of the vertex and of its neighbor. The Markov kernel of our random walk is determined by the following requirement: For any s = 1 , , N , the probability that a random walk starting at some vertex of depth s will reach a leaf before reaching a vertex of depth s 1 equals

(4) q s = ( 2 s W s ) 1 / ( p 1 ) k = s N ( 2 k W k ) 1 / ( p 1 ) .

Thus, the weights of our random walk typically depend on p ( 1 , ) , except for the case where W s is proportional to 2 s . This seems inevitable. Indeed, in some examples such as the 3rd example in Section 2, the linear extension operator H : R V R V that corresponds to the parameter value p = p 0 is not uniformly bounded for any p ( 1 , ) \ { p 0 } . When p = 2 , our random walk coincides with the usual random walk corresponding to the given weights on the edges of the binary tree, and thus in this case, H is the standard harmonic extension operator (hence, C ¯ p = 1 for p = 2 ).

There is a certain range of weights where the averaging operator yields a uniformly bounded linear extension operator, as proven by Björn et al. [1]. The averaging operator is the extension operator that assigns to each internal vertex v the average of the function values on the leaves of the subtree whose root is v . This averaging operator seems natural also from the point of view of Whitney’s extension theory, and see the work by Shvartsman [5] on Sobolev extension in W 1 , p ( R n ) . However, there are examples of radially-symmetric binary trees where the averaging operator does not provide a uniformly bounded operator, such as the case where W k = 2 k for all k .

What about trees with weights that are not radially symmetric? Suppose that the edge weights are arbitrary positive numbers ( W e ) e E that are not necessarily determined by the depth of the edge. When p = 2 , there is still a harmonic extension operator of norm one from W ˙ 1 , p ( V ) to W ˙ 1 , p ( V ) . However, for p 2 , the situation seems subtle. We conjecture that in the general case of nonradially-symmetric tree weights, there is no linear extension operator whose norm is bounded by a function of p alone. This conjecture is closely related to questions about well-complemented subspaces of p n that are beyond the scope of this note.

To prove Theorem 1.1, we reformulate the problem in a way that brings us closer to analysis in p -spaces. For 1 < p < , the associated L p ( E ) -norm is defined, for f : E R , via

(5) f p = f L p ( E ) = k = 1 N W k x { 0 , 1 } k f ( x ) p 1 / p .

We think of a function f : E R as the gradient of a function f ˜ : V R , uniquely determined up to an additive constant. Given f : E R , we thus define a function f ˜ : V R as follows: For any x V other than the root,

(6) f ˜ ( x ) = i = 1 d ( x ) f ( π i x ) ,

while f ˜ ( x ) = 0 if x = is the root. The only property of f ˜ that matters is that for all x E ,

f ( x ) = f ˜ ( x ) f ˜ ( π d ( x ) 1 ( x ) ) .

We are interested in finding a linear operator T : L p ( E ) L p ( E ) , with a uniform bound on its operator norm, which has the following properties:

  1. The operator T takes the form

    (7) T f ( x ) = T ˜ f ˜ ( x ) T ˜ f ˜ ( π d ( x ) 1 x )

    for some linear operator T ˜ : R V R V . That is, T ˜ takes functions on V to functions on V , and T is induced from T ˜ via formula (7).

  2. The operator T ˜ is equivariant with respect to the tree symmetries and it satisfies T ˜ ( 1 ) 1 , i.e., it maps the constant function 1 to itself.

  3. The function T ˜ g coincides with the function g on the leaves of the tree, i.e., ( T ˜ g ) V = g V for any g : V R .

  4. The function T ˜ g is determined by the values of the function g on the leaves of the tree.

The operator norm of T ˜ with respect to the W ˙ 1 , p -seminorm equals to the operator norm of T with respect to the L p ( E ) -norm. Defining H ( f V ) = T ˜ f , Theorem 1.1 may thus be reformulated as follows:

Theorem 1.2

Let 1 < p < and let W 1 , , W N > 0 . Then there exists a linear operator T : L p ( E ) L p ( E ) with the aforementioned properties, whose operator norm is at most a certain constant C ¯ p depending only on p . In fact, we have the bound (3) for the constant C ¯ p .

The proof of Theorem 1.2 occupies the next three sections. In Section 2, we discuss invariant random walks on a full binary tree and describe the corresponding harmonic extension operator. In Section 3, we deal with the problem of bounding the norm of this operator and use the symmetries of the problem to reduce it to a one-dimensional question. This one-dimensional question is then answered in Section 4 using the Muckenhoupt criterion [4].

When analyzing the binary tree, we use the following notation: We write dlca ( x , y ) for the length of the maximal prefix shared by the strings x { 0 , 1 } k and y { 0 , 1 } , while l c a ( x , y ) is the maximal prefix itself. Thus for two vertices x , y V , their least common ancestor is l c a ( x , y ) V and its depth is dlca ( x , y ) { 0 , , N } . Note that for any x E and ω V ,

(8) dlca ( π d ( x ) 1 x , ω ) = min { d ( x ) 1 , dlca ( x , ω ) } .

2 Invariant random walks

A Markov chain on V is a sequence of random variables R 1 , R 2 , V such that the distribution of R i + 1 conditioned on R 1 , , R i is the same as its distribution conditioned on R i . A Markov chain is time-homogeneous if for any x , y V , the probability that R i + 1 = x conditioned on the event R i = y does not depend on i . A random walk on V is a time-homogeneous Markov chain R 1 , R 2 , V such that R i + 1 is a neighbor of R i with probability one.

We say that the random walk is invariant if the probability to jump from a vertex x to a vertex y depends only on the depths d ( x ) and d ( y ) . Our random walk will be invariant, and it will stop when it reaches a leaf, i.e., we have the stopping time

τ = min { i 1 ; R i is a leaf } .

For s 1 , we define q s to be the probability of the following event: Assuming that R 1 is a vertex of depth s , the event is that R i will remain at the subtree whose root is R 1 for all 1 i τ . Equivalently, define

X i = d ( R i ) { 0 , , N } .

Then X 1 , X 2 , { 0 , , N } is a random walk, since X i + 1 X i = 1 for all i . Furthermore,

(9) q s = P ( 1 i τ , X i s X 1 = s ) .

Clearly,

(10) q 0 = q N = 1 .

For r , s { 0 , , N } with r s , we set

p s , r = P ( min { X i ; i τ } = r X 1 = s ) .

That is, the number p s , r is the probability that r is the minimal node that the walker visits when starting from node s , before reaching the terminal node N . Clearly, r = 0 s p s , r = 1 .

Lemma 2.1

For 0 r s N 1 ,

(11) p s , r = q r k = r + 1 s ( 1 q k ) ,

where an empty product equals 1. Moreover, p N , r = δ r N , where δ r N is the Kronecker delta.

Proof

The expression on the right-hand side of (11) is the probability to ever reach s 1 when starting from X 1 = s , and from s 1 to ever reach s 2 , etc. until we finally reach r , yet from r we require to never reach r 1 . Alternatively, when 0 r s , s 1 , we have the recurrence relation:

(12) p s , r = q s δ s , r + ( 1 q s ) p s 1 , r .

This recurrence relation leads to another proof of (11).□

Suppose that our random walk R 1 , R 2 , V begins at a vertex R 1 = x with d ( x ) = s . Consider a leaf y V with dlca ( x , y ) = r . What is the probability that our random walk will reach the leaf y ? We claim that this probability is

(13) b s , r P ( R τ = y ) = k = 0 r 2 k N p s , k .

Indeed, conditioning on the value of k = min i τ d ( R i ) , by symmetry, we know that R τ is distributed uniformly among the 2 N k leaf descendants of the vertex π k ( x ) . When k r , exactly one of these leaf-descendants is the leaf y , since the vertex of minimal depth that ( R i ) visits must be the vertex π k ( x ) , which is a prefix of y as k r = dlca ( x , y ) . Hence, the probability that R τ = y , conditioning on the value of k , equals to 1 / 2 N k when k r , and it vanishes otherwise. By using the definition of p s , k and the complete probability formula, we obtain (13). The harmonic extension operator associated with our invariant random walk is given by

(14) T ˜ g ( x ) = ω { 0 , 1 } N b d ( x ) , dlca ( x , ω ) g ( ω ) ( x V ) .

The operator T is induced from T ˜ via formula (7). Requirements 1 , , 4 from Section 1 are clearly satisfied.

We stipulate that the collection of descendants of a vertex x V , denoted by D ( x ) V , includes the vertex x itself. Abbreviate a b = min { a , b } and a b = max { a , b } . The operator T takes the form

(15) T f ( x ) = y E K ( x , y ) f ( y ) ,

where the kernel K is described next.

Proposition 2.2

Let x , y E and denote s = d ( x ) , t = d ( y ) , r = dlca ( x , y ) . Then the followings hold: If x D ( y ) and y D ( x ) , then r s t 1 and

(16) K ( x , y ) = q s 2 t k = 0 r 2 k p s 1 , k 0 .

Otherwise, i.e., if y D ( x ) or if x D ( y ) , then r = s t and

(17) K ( x , y ) = q s 2 t k = 0 r 1 ( 2 r 2 k ) p s 1 , k 0 .

Proof

By (7) and (14), we have, for any x E ,

(18) T f ( x ) = T ˜ f ˜ ( x ) T ˜ f ˜ ( π d ( x ) 1 x ) = ω { 0 , 1 } N b d ( x ) , dlca ( x , ω ) f ˜ ( ω ) ω { 0 , 1 } N b d ( x ) 1 , dlca ( π d ( x ) 1 x , ω ) f ˜ ( ω ) = ω { 0 , 1 } N a d ( x ) , dlca ( x , w ) f ˜ ( ω ) ,

where for any 0 r s , by (8) and (13),

(19) a s , r = b s , r b s 1 , min { s 1 , r } = k = 0 r 2 k N p s , k k = 0 min { s 1 , r } 2 k N p s 1 , k .

Hence, by (6) and (18),

T f ( x ) = ω { 0 , 1 } N a d ( x ) , dlca ( x , ω ) i = 1 N f ( π i ω ) = y E ω { 0 , 1 } N ; π d ( y ) ω = y a d ( x ) , dlca ( x , ω ) f ( y ) = y E K ( x , y ) f ( y ) ,

where the kernel K of the operator T satisfies, for any x , y E ,

(20) K ( x , y ) = ω { 0 , 1 } N ; π d ( y ) ω = y a d ( x ) , dlca ( x , ω ) .

Fix x , y E with s = d ( x ) , t = d ( y ) and r = dlca ( x , y ) . Let us consider first the case where x is not a descendant of y . This means that the prefix of y that is shared by x , is not the entire string y . Hence, for any ω { 0 , 1 } N with π d ( y ) ω = y , we have dlca ( x , ω ) = dlca ( x , y ) d ( y ) 1 . Therefore, from (20) and (19),

K ( x , y ) = 2 N d ( y ) a d ( x ) , dlca ( x , y ) = 2 N t [ b s , r b s 1 , ( s 1 ) r ] = k = 0 r 2 k t p s , k k = 0 ( s 1 ) r 2 k t p s 1 , k ,

as b s , r = k = 0 r 2 k N p s , k . Since p s , s = q s , we have

(21) K ( x , y ) = δ r s 2 s t q s + k = 0 ( s 1 ) r 2 k t [ p s , k p s 1 , k ] = q s δ r s 2 s t k = 0 ( s 1 ) r 2 k t p s 1 , k ,

where we used relation (12), which implies that when k s 1 ,

(22) p s , k p s 1 , k = q s p s 1 , k .

We may now prove the conclusion of the proposition in the case where x D ( y ) and y D ( x ) . Indeed, in this case, r s t 1 and formula (21) applies. Since δ r s = 0 in this case, we deduce formula (16) from (21).

The next case we consider is the case where r = s t 1 , or equivalently, where y D ( x ) \ { x } . Thus, x D ( y ) and formula (21) applies. Recalling that k = 0 s 1 p s 1 , k = 1 , we obtain from (21) that

K ( x , y ) = q s k = 0 s 1 ( 2 s t 2 k t ) p s 1 , k = q s 2 t k = 0 s 1 ( 2 r 2 k ) p s 1 , k ,

proving formula (17) in the case y D ( x ) \ { x } .

We move on to the case where x D ( y ) \ { y } , thus dlca ( x , y ) = t s 1 . In this case, by applying (20), (19), (13) and then (22),

K ( x , y ) = ω { 0 , 1 } N ; π d ( y ) ω = y a d ( x ) , dlca ( x , ω ) = 2 N d ( x ) a d ( x ) , d ( x ) + k = d ( y ) d ( x ) 1 2 N k 1 a d ( x ) , k = 2 N s a s , s + k = t s 1 2 N k 1 a s , k = 2 N s ( b s , s b s 1 , s 1 ) + k = t s 1 2 N k 1 ( b s , k b s 1 , k ) = 2 N s k = 0 s 2 k N p s , k k = 0 s 1 2 k N p s 1 , k + k = t s 1 2 N k 1 = 0 k 2 N ( p s , p s 1 , ) = q s q s k = 0 s 1 2 k s p s 1 , k q s = 0 s 1 k = t s 1 2 k 1 p s 1 , = q s 1 k = 0 s 1 2 k s p s 1 , k = 0 s 1 [ 2 t 2 s ] p s 1 , = q s 1 k = 0 s 1 2 k k t p s 1 , k = q s 1 k = 0 t 1 2 k t p s 1 , k k = t s 1 p s 1 , k = q s k = 0 t 1 ( 1 2 k t ) p s 1 , k = q s 2 t k = 0 t 1 ( 2 t 2 k ) p s 1 , k .

Since r = t in this case, we have proved formula (17) in the case where y D ( x ) \ { x } . Finally, the last case that remains is when x = y . In this case, r = s = t and

K ( x , y ) = ω { 0 , 1 } N ; π d ( y ) ω = y a d ( x ) , dlca ( x , ω ) = 2 N d ( x ) a d ( x ) , d ( x ) = 2 N s a s , s = 2 N s ( b s , s b s 1 , s 1 ) = 2 N s k = 0 s 2 k N p s , k k = 0 s 1 2 k N p s 1 , k = q s q s k = 0 s 1 2 k s p s 1 , k = q s k = 0 s 1 ( 1 2 k s ) p s 1 , k ,

completing the proof of formula (17).□

Some examples.

  1. The simplest example is when q s = 1 for all s 1 . In this case, the operator T ˜ is the familiar averaging operator. That is, the extension operator T ˜ is the operator that assigns to each vertex the average of the values at the leaves of its subtree. In this case,

    p s , r = δ s , r .

  2. Consider the case where the invariant random walk is such that X i d ( R i ) is a symmetric random walk on { 0 , , N } , i.e., the probability to jump from i to i + 1 is exactly 1/2 for i = 1 , , N 1 . Recall that q s is the probability to never leave the subtree when starting at a vertex of depth s . We claim that in this example, for s = 1 , , N ,

    (23) q s = 1 N s + 1 .

    Indeed, the function f ( i ) = i is harmonic on { 0 , 1 , , N } , and hence, f ( X i ) is a martingale. Thus, for any stopping time τ ˜ , we have f ( X 1 ) = E f ( X τ ˜ ) . We pick the stopping time

    τ ˜ = min { i ; X i { s 1 , N } }

    and obtain (23) since

    N q s + ( s 1 ) ( 1 q s ) = s .

    Next, we use Lemma 2.1 and find a formula for p s , r . Since formula (23) is valid for any s 1 , we conclude that for any r 0 and s r + 1 ,

    (24) k = r + 1 s ( 1 q k ) = k = r + 1 s N k N k + 1 = N s N r .

    Formula (24) is actually valid for any 0 r s , since an empty product equals one. Recall that q 0 = 1 . We thus conclude from Lemma 2.1 that for s = 0 , , N 1 ,

    p s , r = N s N r 1 r = 0 1 N r + 1 1 r s ,

    while p N , r = δ N , r .

  3. Let 0 < δ < 1 , and consider the case where ( X i ) is a random walk on { 0 , , N } such that the probability to jump from k to k + 1 equals 1/2 if k < N 1 , and it equals δ if k = N 1 . A harmonic function here is

    f ( k ) = k k N 1 N 2 + 1 / δ k = N .

    Therefore, for s = 1 , , N 1 ,

    ( N 2 + 1 / δ ) q s + ( s 1 ) ( 1 q s ) = s .

    Thus, q 0 = q N = 1 , while for s = 1 , , N 1 ,

    q s = 1 N s + 1 / δ 1 .

    Hence, for any s N 1 and r s 1 ,

    k = r + 1 s ( 1 q k ) = k = r + 1 s N k + δ 1 2 N k + δ 1 1 = N s + δ 1 2 N r + δ 1 2 .

    We conclude from Lemma 2.1 that for s = 0 , , N 1 and 0 r s ,

    p s , r = q r k = r + 1 s ( 1 q k ) = N s + δ 1 2 N r + δ 1 2 1 r = 0 1 N r + 1 / δ 1 1 r s ,

    while p N , r = δ N , r .

We conclude this section with the following:

Lemma 2.3

For any numbers q 1 , , q N 1 ( 0 , 1 ) , there exists a random walk

X 1 , X 2 , { 0 , , N }

satisfying (9) with τ = min { i 1 ; X i = N } for s = 1 , , N 1 .

Proof

Write x k for the probability that the random walk jumps from k to k + 1 . Then for s = 0 , , N 1 ,

(25) q s = x s ( q s + 1 + ( 1 q s + 1 ) x s ) ,

where we set q 0 = q N = 1 . We claim that x s [ 0 , 1 ] is determined by equation (25). Indeed, the right-hand side of (25) is a continuous, increasing function of x s in the interval [ 0 , 1 ] , which maps this interval to itself.□

3 The ancestral and nonancestral parts of the kernel

We need to bound the operator norm in L p ( E ) of the operator T whose kernel is described in Proposition 2.2. Let us consider first the nonancestral part of the operator, given by the kernel

(26) K 0 ( x , y ) = K ( x , y ) 1 { x D ( y ) , y D ( x ) } = 1 { r s t 1 } q s 2 t k = 0 r 2 k p s 1 , k .

Here, as usual s = d ( x ) , t = d ( y ) , and r = dlca ( x , y ) . Write T 0 : L p ( E ) L p ( E ) for the operator whose kernel is K 0 . A function f : E R is invariant under the symmetries of the tree, or invariant in short, if it takes the form

f ( x ) = F ( d ( x ) )

for some function F : { 1 , , N } R . The operator T 0 is equivariant under the symmetries of the tree. Therefore, if f ( x ) = F ( d ( x ) ) is an invariant function, then so is T 0 f . In fact, in the case where f ( x ) = F ( d ( x ) ) , we can write

(27) T 0 f ( x ) = t = 1 N L 0 ( d ( x ) , t ) F ( t )

for a certain kernel L 0 ( s , t ) defined for s , t = 1 , , N .

Lemma 3.1

For s , t = 1 , , N ,

L 0 ( s , t ) = q s k = 0 m 1 ( 1 2 k m ) p s 1 , k 0 .

Proof

Let r min { t , s } 1 . A moment of reflection reveals that for x E with d ( x ) = s ,

n ( t ; s , r ) # { y E ; d ( y ) = t , dlca ( x , y ) = r } = 2 t r 1 .

By (26), (27), and the definition of T 0 , for any x E with d ( x ) = s ,

L 0 ( s , t ) = y E ; d ( y ) = t K 0 ( x , y ) = r = 0 t s 1 n ( t ; s , r ) q s 2 t k = 0 r 2 k p s 1 , k .

Denote m = s t . Then for s , t = 1 , , N ,

L 0 ( s , t ) = q s r = 0 t s 1 k = 0 r 2 k r 1 p s 1 , k = q s k = 0 m 1 r = k m 1 2 k r 1 p s 1 , k = q s k = 0 m 1 ( 1 2 k m ) p s 1 , k .

Write Ω N = { 1 , , N } and for F : Ω N R define

(28) F p = F L p ( Ω N ) = k = 1 N 2 k W k F ( k ) p 1 / p .

Observe that f L p ( E ) = F p if f ( x ) = F ( d ( x ) ) . Let

S 0 F ( s ) = t = 1 N L 0 ( s , t ) F ( t )

so that by (27),

(29) T 0 ( F d ) = ( S 0 F ) d .

For f , g : E R , we consider the scalar product

f , g = x E W d ( x ) f ( x ) g ( x ) ,

while for F , G : Ω N R , we set

F , G F d , G d = k = 1 N 2 k W k F ( k ) G ( k ) .

The adjoint operators T 0 and S 0 are defined with respect to these scalar products. The following lemma is probably well known to experts (see, e.g., Howard and Schep [2] for a related argument), and its proof is provided for completeness.

Lemma 3.2

Let 1 < p < . Then the norm of the operator T 0 : L p ( E ) L p ( E ) is attained at an invariant, nonnegative function f, and it equals to the norm of the operator S 0 : L p ( Ω n ) L p ( Ω n ) .

Proof

Denote momentarily T = T 0 and S = S 0 . By Lemma 3.1, the kernel L 0 of the operator S is nonnegative, and by (26), the kernel of the operator T is nonnegative as well. By approximation, we may assume that these two kernels are strictly positive, while keeping condition (29), thus

(30) T ( F d ) = ( S F ) d .

We deduce that T ( F d ) = ( S F ) d . By compactness,

sup 0 F L p ( Ω N ) S F p F p

is attained at some function F . Since the kernel of S is nonnegative, we may assume that the extremal function F is nonnegative. By the Lagrange multipliers theorem, the function F satisfies a certain eigenvalue equation, and in fact, there exists λ R such that

(31) S ( S F ) p 1 = λ F p 1 .

Since the kernel of S is positive and F is nonnegative and not identically zero, it follows from (31) that F is actually positive. The norm of the operator S : L p ( Ω n ) L p ( Ω n ) equals λ 1 / p > 0 , since

S F p p = ( S F ) p 1 , S F = S ( S F ) p 1 , F = λ F p 1 , F = λ F p p .

Denoting f = F d , we find from (30) that f is a positive invariant function satisfying

T ( T f ) p 1 = λ f p 1 .

Since the kernel of T is nonnegative, we have the pointwise Hölder inequality

T ( u v ) T ( u p ) 1 / p T ( v q ) 1 / q ,

valid for any nonnegative functions u , v L p ( E ) , where q = p / ( p 1 ) . The operator T has a nonnegative kernel, and hence, its norm is attained at a nonnegative function u L p ( E ) . By the pointwise Hölder inequality,

( T u ) p T ( u p f p / q ) ( T f ) p / q = T ( u p f 1 p ) ( T f ) p 1 .

Therefore,

T u L p ( E ) p T ( u p f 1 p ) , ( T f ) p 1 = u p f 1 p , T ( T f ) p 1 = λ u p f 1 p , f p 1 = λ u L p ( E ) p .

Thus, the norm of T : L p ( E ) L p ( E ) is at most λ 1 / p , which is the norm of S : L p ( Ω N ) L p ( Ω N ) . The two norms must therefore be equal, since the operator S is equivalent to the restriction of T to the space of invariant functions.□

We move on to the ancestral part of the operator, which according to Proposition 2.2 is given by

(32) K 1 ( x , y ) = K ( x , y ) K 0 ( x , y ) = 1 { r = s t } q s 2 t k = 0 r 1 ( 2 r 2 k ) p s 1 , k 0 .

Write T 1 for the operator whose kernel is K 1 . As before, for an invariant function f ( x ) = F ( d ( x ) ) , we may write

(33) T 1 f ( x ) = t = 1 N L 1 ( d ( x ) , t ) F ( t )

for a certain kernel L 1 ( s , t ) defined for s , t = 1 , , N . We also write

S 1 F ( s ) = t = 1 N L 1 ( s , t ) F ( t ) .

It is possible to use formula (7) for the operator T and the definition (14) of the harmonic extension operator T ˜ and deduce that

(34) S 0 + S 1 0 ,

essentially because the only invariant, harmonic function on the vertices of the tree is the constant function. An alternative, more direct proof of (34) is provided in the following:

Lemma 3.3

Let 1 < p < . Then the norm of the operator T 1 : L p ( E ) L p ( E ) is equal to the norm of S 1 : L p ( Ω n ) L p ( Ω n ) . In addition, for s , t = 1 , , N with m = s t , we have

L 1 ( s , t ) = q s k = 0 m 1 ( 1 2 k m ) p s 1 , k = L 0 ( s , t ) .

Proof

The first assertion of the lemma follows from the fact that the kernel of T 1 is nonnegative and invariant under the symmetries of the tree, as in Lemma 3.2. For the second part, let s , t = 1 , , N and denote m = s t . We claim that for x E with d ( x ) = s ,

(35) n ( t ; s , m ) # { y E ; d ( y ) = t , dlca ( x , y ) = m } = max { 1 , 2 t s } .

Indeed, assume first that t s . How many y s are there with d ( y ) = t and dlca ( x , y ) = m ? Since m = s , the answer is 2 t s . Next, if s t , then the number of such y s is one. This proves (35). Therefore,

L 1 ( s , t ) = y E ; d ( y ) = t K 1 ( x , y ) = n ( t ; s , m ) q s 2 t k = 0 m 1 ( 2 m 2 k ) p s 1 , k = max { 2 t , 2 s } q s k = 0 m 1 ( 2 m 2 k ) p s 1 , k = 2 m q s k = 0 m 1 ( 2 m 2 k ) p s 1 , k .

Corollary 3.4

We have

T L p ( E ) L p ( E ) 2 S 0 L p ( Ω N ) L p ( Ω N ) .

Proof

This follows from the fact that T = T 0 + T 1 together with the facts that T 0 = S 0 and T 1 = S 1 , while S 1 = S 0 .□

In view of Corollary 3.4, we are interested in bounds for the norm of the operator S = S 0 = S 1 : L p ( Ω N ) L p ( Ω N ) whose nonnegative kernel is

(36) L ( s , t ) = q s k = 0 s t 1 ( 1 2 k s t ) p s 1 , k q s k = 0 s t 1 p s 1 , k .

From Lemma 2.1, we know that p s , r = q r k = r + 1 s ( 1 q k ) for s N 1 . A little exercise in probability shows that for any 1 m s N ,

(37) k = 0 m 1 p s 1 , k = k = m s 1 ( 1 q k ) .

Alternatively, (37) holds true for m = 0 as q 0 = 1 , and it may be proven by induction on m since

p s 1 , m + k = m s 1 ( 1 q k ) = q m k = m + 1 s 1 ( 1 q k ) + k = m s 1 ( 1 q k ) = k = m + 1 s ( 1 q k ) .

From (36) and (37), we thus obtain

Corollary 3.5

For s , t = 1 , , N , with m = min { s , t } ,

0 L ( s , t ) q s k = m s 1 ( 1 q k ) ,

where an empty product equals one.

Some examples (parallel to the ones discussed in Section 2).

  1. For the averaging operator, where q s = 1 and p s , r = δ s , r , we have

    L ( s , t ) = 1 2 1 { s t } ,

    i.e., this is the matrix whose entries equal 0 below the diagonal and 1 / 2 on and above the diagonal. This is a rather simple matrix, and it is bounded with respect to the weighted L p -norm for quite a few sequences of weights.

  2. For the symmetric random walk matrix, we have q 0 = 1 while for 1 s N ,

    q s = 1 N s + 1 .

    Hence, in view of Corollary 3.5, with m = min { s , t } ,

    0 L ( s , t ) 1 N s + 1 k = m s 1 N k N k + 1 = 1 N m + 1 .

  3. In the case where

    q s = 1 N s + 1 / δ 1

    for some 0 < δ < 1 , we have

    0 L ( s , t ) 1 N s + 1 / δ 1 k = m s 1 N k + 1 / δ 2 N k + 1 / δ 1 = 1 N m + 1 / δ 1 .

All that remains is to bound the L p ( Ω N ) -norm of the operator whose kernel is discussed in Corollary 3.5. Recall from Lemma 2.3 that we have the freedom to choose the parameters q 1 , , q N 1 ( 0 , 1 ) as we please.

How should we choose these parameters? Since L ( N , t ) k = t N 1 ( 1 q k ) and we are looking for upper bounds for the norm, the q s should not be too tiny. On the other hand, L ( s , t ) q s for s N 1 , and hence, it is beneficial to choose q s rather small. We would therefore need some balance for the q s , which is the subject of the next section.

4 One-dimensional analysis

Let 1 < p < . It will be slightly more convenient to denote

K ( s , t ) = L ( N + 1 s , N + 1 t ) and Q s = q N + 1 s .

Recalling from (10) that q N = 1 , we see that

Q 1 = 1 .

From Corollary 3.5, we know that for s , t = 1 , , N ,

K ( s , t ) = L ( N + 1 s , N + 1 t ) q N + 1 s k = N + 1 max { s , t } N s ( 1 q k ) = Q s k = s + 1 max { s , t } ( 1 Q k ) .

From Corollary 3.5, we know that L 0 . Consequently, for s , t = 1 , , N ,

(38) 0 K ( s , t ) Q s t s Q s k = s + 1 t ( 1 Q k ) t s + 1 ,

Recall that we are given edge weights W 1 , , W N > 0 , and that the associated L p ( E ) -norm is given by (5). Denote

w s 2 N + 1 s W N + 1 s > 0 .

Consider the weighted p -norm

(39) f p , w = k = 1 N w k f ( k ) p 1 / p ,

and the operator T whose kernel is K ( s , t ) . We are allowed to choose the weights q 1 , , q N 1 ( 0 , 1 ) as we please, or equivalently, we have the freedom to determine Q 2 , , Q N ( 0 , 1 ) . We must keep Q 1 = 1 . On the basis of the considerations related to the Muckenhoupt criterion discussed below, we set

(40) Q s = w s 1 / ( p 1 ) k = 1 s w k 1 / ( p 1 ) .

It is clear that Q 1 = 1 and that Q s ( 0 , 1 ) for all s 2 . Recall that q = p / ( p 1 ) .

Lemma 4.1

To prove Theorem 1.2, it suffices to show that the operator norm of T with respect to the p , w -norm is bounded by a constant C ˆ p > 0 depending only on p, where in fact

(41) C ˆ p 2 p 1 / p q 1 / q ( 1 + max { ( p 1 ) 1 / p , ( q 1 ) 1 / q } ) .

Proof

In view of Corollary 3.4, it suffices to bound the operator norm of S 0 , whose kernel is L , with respect to the L p ( Ω N ) -norm defined in (28). Under the transformation

s N + 1 s ,

the operator S 0 whose kernel is L transforms to the operator T whose kernel is K . The L p ( Ω N ) -norm from (28) transforms to the p , w -norm defined in (39). Hence, Theorem 1.2 would follow once we obtain the bound (41), where C ¯ p 2 C ˆ p by Corollary 3.4.□

The remainder of this section is devoted to the proof of the following:

Proposition 4.2

The operator norm of T with respect to the norm (39) is bounded by a number C ˆ p depending only on p ( 1 , ) . In fact, we have the bound (41) for the constant C ˆ p .

Our main tool in the proof of Proposition 4.2 is the Muckenhoupt criterion [4], which is an indispensable tool for proving one-dimensional inequalities of the Poincaré-Sobolev type. For the reader’s convenience, we include here a statement and a proof of a straightforward modification of the Muckenhoupt criterion, with sums in place of integrals:

Theorem 4.3

(Muckenhoupt) Let 1 < p < and write Ω N = { 1 , , N } . Let U , V : Ω N ( 0 , ) and let A > 0 be such that for all r = 1 , , N ,

(42) k = r N U ( k ) p 1 / p A k = 1 r V ( k ) q 1 / q .

Then for any function f : Ω N R ,

(43) k = 1 N U ( k ) = 1 k f ( ) p 1 / p C p A k = 1 N V ( k ) f ( k ) p 1 / p ,

with C p = p 1 / p q 1 / q .

By continuity, the analog of Theorem 4.3 for p = 1 , holds true with C 1 = C = 1 . We remark that as in [4], this criterion is tight, in the sense that the infimum over all A > 0 satisfying (42) is equivalent to the best constant in inequality (43). For the proof of Theorem 4.3, we require the following:

Lemma 4.4

For 1 < p < , q = p / ( p 1 ) , α 1 , , α N > 0 and r = 1 , , N ,

(44) k = 1 r α k = 1 k α 1 / p q k = 1 r α k 1 / q .

Proof

We will use the simple inequality

(45) ( a + b ) 1 / q a 1 / q b min ξ ( a , a + b ) ξ 1 / q 1 q = 1 q ( a + b ) 1 / q 1 b ,

valid for any a , b 0 with a + b > 0 . From (45), for k = 1 , , N ,

= 1 k α 1 / q = 1 k 1 α 1 / q 1 q α k = 1 k α 1 / q 1 ,

where an empty sum equals zero. By summing this for k = 1 , , r , we obtain (44).□

Proof of Theorem 4.3

(Muckenhoupt) By the Hölder inequality, for any function f : Ω N R and weights h : Ω N ( 0 , ) ,

(46) k = 1 N U ( k ) = 1 k f ( ) p k = 1 N U p ( k ) = 1 k f ( ) V ( ) h ( ) p j = 1 k V ( j ) h ( j ) q p / q = = 1 N f ( ) V ( ) h ( ) p k = N U p ( k ) j = 1 k V ( j ) h ( j ) q p / q .

Set h ( k ) = = 1 k V ( ) q 1 / ( p q ) . By applying Lemma 4.4 with α k = V ( k ) q , we obtain

k = 1 r V ( k ) h ( k ) q = k = 1 r α k = 1 k α 1 / p q k = 1 r α k 1 / q = q k = 1 r V ( k ) q 1 / q .

Hence, for any f : Ω N R , the expression in (46) is at most

(47) q p / q = 1 N f ( ) V ( ) h ( ) p k = N U p ( k ) j = 1 k V ( j ) q p / q 2 .

By applying (42) and then Lemma 4.4 with α k = U p ( N + 1 k ) and with p ( 1 , ) playing the rôle of q ( 1 , ) , we see that

k = N U ( k ) p j = 1 k V ( j ) q p / q 2 A p / q k = N U ( k ) p j = k N U ( j ) p 1 / q = A p / q k = 1 N + 1 α k j = 1 k α j 1 / q p A p / q k = 1 N + 1 α k 1 / p = p A p / q k = N U p ( k ) 1 / p .

Hence, the expression in (47) is at most

p ( q A ) p / q = 1 N f ( ) V ( ) h ( ) p k = N U p ( k ) 1 / p .

By applying (42) again, we bound the last expression from above by

p ( q A ) p / q A = 1 N f ( ) V ( ) h ( ) p k = 1 V q ( k ) 1 / q = p q p / q A p = 1 N f ( ) V ( ) p ,

hence completing the proof.□

Corollary 4.5

Let 1 < p < and Ω N = { 1 , , N } . Let U , V : Ω N ( 0 , ) and let A > 0 be such that for all r = 1 , , N ,

(48) k = 1 r U ( k ) p 1 / p A k = r N V ( k ) q 1 / q .

Then for any f : Ω N R ,

(49) k = 1 N U ( k ) = k N f ( ) p 1 / p C p A k = 1 N V ( k ) f ( k ) p 1 / p ,

with C p = p 1 / p q 1 / q .

Proof

Denote U ˜ ( k ) = U ( N + 1 k ) and V ˜ ( k ) = V ( N + 1 k ) . Then from (48), for all r = 1 , , N ,

k = r N U ˜ ( k ) p 1 / p A k = 1 r V ˜ ( k ) q 1 / q .

By Theorem 4.3, this implies that for any g : Ω N R , denoting f ( r ) = g ( N + 1 r ) ,

k = 1 N U ˜ ( k ) = 1 k f ( N + 1 ) p 1 / p C p A k = 1 N V ˜ ( k ) f ( N + 1 k ) p 1 / p ,

or equivalently,

k = 1 N U ˜ ( N + 1 k ) = k N f ( ) p 1 / p C p A k = 1 N V ˜ ( N + 1 k ) f ( k ) p 1 / p .

This implies (49).□

Proposition 4.6

For f : Ω n R and s = 1 , , N denote

T 0 f ( s ) = Q s t = 1 s f ( t ) ,

where Q s is defined in (40). Then the operator norm of T 0 with respect to the norm p , w defined in (39) is bounded by a number C ˜ p depending only on p ( 1 , ) . In fact, C ˜ p 2 1 / p p 1 / p q 1 / q max { 1 , ( p 1 ) 1 / p } .

The proof of Proposition 4.6 requires the following:

Lemma 4.7

For 1 < p < , α 1 , , α N > 0 and r = 1 , , N ,

(50) k = r N α k = 1 k α p 2 min { 1 , p 1 } k = 1 r α k p + 1 .

Proof

We use the inequality

( p 1 ) b ( a + b ) p a p + 1 ( a + b ) p + 1 ,

which is valid for any a , b > 0 . Then for k = 2 , , N ,

( p 1 ) α k = 1 k α p = 1 k 1 α p + 1 = 1 k α p + 1 .

By summing this for k = r + 1 , , N , we obtain

( p 1 ) k = r + 1 N α k = 1 k α p k = 1 r α k p + 1 = 1 N α p + 1 k = 1 r α k p + 1 ,

where an empty sum equals zero. We conclude (50) by summing this with the trivial inequality

min { 1 , p 1 } α r = 1 r α p k = 1 r α k p + 1 .

Proof of Proposition 4.6

Define

U ( k ) = w k 1 / p Q k and V ( k ) = w k 1 / p .

Let us verify the condition of the Muckenhoupt criterion. We need to find A > 0 such that for all r = 1 , , N inequality (42) holds true, that is,

(51) k = r N w k Q k p A p k = 1 r w k q / p p / q .

Recall that p / q = p 1 . From the definition (40) of Q k , we need

k = r N w k 1 / ( p 1 ) = 1 k w 1 / ( p 1 ) p A p k = 1 r w k 1 / ( p 1 ) ( p 1 ) .

By setting α k = w k 1 / ( p 1 ) and using Lemma 4.7, we see that (51) holds true with

A = 2 1 / p max { 1 , ( p 1 ) 1 / p } .

From Theorem 4.3, we thus conclude that for any f : Ω N R ,

k = 1 N w k Q k = 1 k f ( ) p 1 / p p 1 / p q 1 / q A k = 1 N w k f ( k ) p 1 / p .

This implies the required bound for the operator norm of f .□

Proposition 4.8

For f : Ω n R and s = 1 , , N denote

T 1 f ( s ) = t = s + 1 N Q s k = s + 1 t ( 1 Q k ) f ( t ) ,

where Q s is as in (40). Then the operator norm of T 1 with respect to the norm p , w defined in (39)is bounded by 2 1 / q p 1 / p q 1 / q max { 1 , ( q 1 ) 1 / q } .

Proof

Denote α k = w k 1 / ( p 1 ) and recall from (40) that Q k = α k / = 1 k α . We claim that for any t s + 1 ,

(52) Q s k = s + 1 t ( 1 Q k ) = α s k = 1 t α k .

Indeed, (52) holds true for t = s , since an empty product equals one, and for t s + 1 , it is proven by an easy induction on t . Consequently,

T 1 f ( s ) = α s t = s + 1 N 1 j = 1 t α j f ( t ) .

Since p , q ( 1 , ) , the elementary inequality of Lemma 4.7 is valid also when p is replaced by q . It implies that for r = 1 , , N ,

(53) k = 1 r α k q + 1 min { 1 , q 1 } 2 k = r N α k j = 1 k α j q .

Set A = 2 1 / q min { 1 , q 1 } 1 / q . Since α k = w k 1 / ( p 1 ) and q / p = q 1 = 1 / ( p 1 ) , it follows from (53) that for r = 1 , , N ,

k = 1 r w k α k p 1 / p A k = r N w k q / p j = 1 k α j q 1 / q .

This is precisely the Muckenhoupt criterion from Corollary 4.5, with

U ( k ) = w k 1 / p α k and V ( k ) = w k 1 / p j = 1 k α j .

Thus, by Corollary 4.5, for any g : Ω N R ,

(54) k = 1 N w k α k = k N g ( ) p ( C p A ) p k = 1 N w k j = 1 k α j g ( k ) p ,

with C p = p 1 / p q 1 / q . By restricting attention to nonnegative functions g in (54), we may alter (54) and replace = k N by the shorter sum = k + 1 N . Inequality (54) remains correct, for nonnegative g , also after this modification. By denoting g ( k ) = f ( k ) / j = 1 k α j , we conclude that for any nonnegative function f : Ω N R ,

(55) k = 1 N w k α k = k + 1 N 1 j = 1 α j f ( ) p ( C p A ) p k = 1 N w k f ( k ) p .

Since the kernel of T 1 is nonnegative, its operator norm is attained at a nonnegative function f : Ω N R . Therefore, (55) implies the required bound for the operator norm of T 1 .□

Proof of Proposition 4.2

The kernel of the operator T is given in (38). It is a nonnegative kernel, and therefore, the operator norm of T is at most the operator norm of the operator whose kernel is the expression on the right-hand side of (38). The latter operator equals

T 0 + T 1

with T 0 from Proposition 4.6 and T 1 from Proposition 4.8. From these two propositions, it follows that the operator norm of T is at most

2 1 / p p 1 / p q 1 / q max { 1 , ( p 1 ) 1 / p } + 2 1 / q p 1 / p q 1 / q max { 1 , ( q 1 ) 1 / q } .

Theorem 1.2 follows from Lemma 4.1 and Proposition 4.2.

Remark

  1. In this article, we have left open several natural questions, including the existence of linear extension operators for W ˙ 1 , p ( T ) for weighted trees T in the extreme cases p = 1 , p = , as well as the analog of our result for the inhomogeneous Sobolev space W 1 , p ( T ) in place of W ˙ 1 , p ( T ) .

  2. The problem of existence of linear Sobolev extension operators for weighted trees arose in connection with an extension problem for W 2 , p ( R 2 ) . More precisely, given E R 2 , let W ˙ 2 , p ( E ) denote the space of restrictions to E of functions in W ˙ 2 , p ( R 2 ) , endowed with the natural seminorm. Does there exist a linear extension operator from W ˙ 2 , p ( E ) to W ˙ 2 , p ( R 2 ) ? The answer is affirmative for p > 2 ; see the study by Israel [3]. For 1 < p < 2 , the answer is unknown. For a particular class of examples E , the problem reduces to the question answered by Theorem 1.1.


Dedicated in friendship to David Jerison.


Acknowledgements

We would like to thank Jacob Carruth, Arie Israel, Anna Skorobogatova and Ignacio Uriarte-Tuero for helpful conversations. This research was conducted while BK was visiting Princeton University’s Department of Mathematics; he is grateful for their gracious hospitality.

  1. Funding information: Charles Fefferman was supported by the Air Force Office of Scientific Research (Grant number FA9950-18-1-0069) and the National Science Foundation (NSF) (Grant number DMS-1700180). Bo’az Klartag was supported by a grant from the Israel Science Foundation (ISF).

  2. Conflict of interest: The authors state no conflict of interest.

References

[1] A. Björn, J. Björn, J. T. Gill, and N. Shanmugalingam, Geometric analysis on Cantor sets and trees, J. Reine Angew. Math. 725 (2017), 63–114. 10.1515/crelle-2014-0099Search in Google Scholar

[2] R. Howard and A. R. Schep, Norms of positive operators on Lp-spaces, Proc. Amer. Math. Soc. 109 (1990), no. 1, 135–146. 10.1090/S0002-9939-1990-1000156-XSearch in Google Scholar

[3] A. Israel, A bounded linear extension operator for L2,p(R2), Ann. Math. two, 178 (2013), no. 1, 183–230. 10.4007/annals.2013.178.1.3Search in Google Scholar

[4] B. Muckenhoupt, Hardyas inequality with weights, Studia Math. 44 (1972), 31–38. 10.4064/sm-44-1-31-38Search in Google Scholar

[5] P. Shvartsman, Sobolev Wp1 -spaces on closed subsets of Rn, Adv. Math. 220 (2009), no. 6, 1842–1922. 10.1016/j.aim.2008.09.020Search in Google Scholar

Received: 2023-02-20
Accepted: 2023-05-17
Published Online: 2023-06-26

© 2023 the author(s), published by De Gruyter

This work is licensed under the Creative Commons Attribution 4.0 International License.

Downloaded on 27.4.2024 from https://www.degruyter.com/document/doi/10.1515/ans-2022-0075/html
Scroll to top button