Harald Fripertinger1,       Ludwig Reich

On covariant embeddings of a linear functional equation with respect to an analytic iteration group

On covariant embeddings of a linear functional equation with respect to an analytic iteration group

Institut für Mathematik
Karl Franzens Universität Graz
Heinrichstr. 36/4
A-8010 Graz, AUSTRIA
Email: harald.fripertinger@kfunigraz.ac.at
Email: ludwig.reich@kfunigraz.ac.at

Abstract

Let a(x), b(x), p(x) be formal power series in the indeterminate x over C (i.e. elements of the ring C [[x]] of such series), such that ord a(x) = 0, ord p(x) = 1 and p(x) is embeddable into an analytic iteration group (p(s,x))s Î C in C [[x]]. By a covariant embedding of the linear functional equation


j(p(x)) = a(x)j(x)+b (x),
(L)
(for the unknown series j(x) Î C [[x]]) with respect to (p(s,x))s Î C we understand families (a(s,x))s Î C and (b(s,x))s Î C with entire coefficient functions in s, such that the system of functional equations and boundary conditions


j(p(s,x)) = a(s,x)j(x)+b(s,x)
(Ls)


a(t+s,x) = a(s,x)a(t,p(s,x))
(Co1)


b(t+s,x) = b(s,x)a(t,p(s,x)) +b(t,p(s,x))
(Co2)


a(0,x) = 1        b(0,x) = 0
(B1)


a(1,x) = a(x)        b(1,x) = b(x)
(B2)
holds for all solutions j(x) of (L) and s,t Î C. In this paper we solve the system ((Co1),(Co2)) (of so called cocycle equations) completely, describe when and how the boundary conditions (B1) and (B2) can be satisfied and present a large class of equations (L) together with iteration groups (p(s,x))s Î C for which there exist covariant embeddings of (L) with respect to (p(s,x))s Î C.

1  Introduction

Let C [[x]] be the ring of formal power series in the indeterminate x with complex coefficients. Consider the linear functional equation


j(p(x)) = a(x)j(x)+b (x),
(L)
where p(x),a(x),b(x) Î C [[x]] are given formal power series and j(x) Î C [[x]] should be determined by the functional equation. We always assume that


p(x) = rx+c2x2+c3x3+... = rx+
å
n ³ 2 
cnxn
with multiplier r ¹ 0 and


a(x) = a0+a1x+a2x2+... =
å
n ³ 0 
anxn
with a0 ¹ 0. For a foundation of the basic calculations with formal power series we refer the reader to [Henrici, 1977] and to [Cartan, 1963] or [Cartan, 1966]. If y(x) Î C [[x]] is of the form y(x) = ån ³ kdkxk with dk ¹ 0, then k is the order of y, which will be indicated as ord y(x) = k. Hence ord p(x) = 1 and ord a(x) = 0. The set of all formal power series of order 1 is indicated by G, which is a group with respect to the substitution in C [[x]]. In addition to this let G0 indicate the set of all formal power series of the form x+d2x2+... Î G.

Furthermore, the notion of congruence modulo order r will be useful. We write j º y mod ord r for formal power series j(x),y(x) Î C [[x]] if the difference j(x)-y(x) is a series of order greater than or equal to r.

A formal power series j(x) can be substituted into the series y(x) = ån ³ 0dnxn Î C [[x]], i.e. the series y(j(x)) = ån ³ 0dnj(x)n can be computed, if and only if ord j(x) ³ 1.

The exponential series is given as


exp(x) =
å
n ³ 0 
xn
n!
,
and the formal logarithm is the series defined by


ln(1+x) =
å
n ³ 1 
(-1)nxn
n
.

A family p: = (p(s,·))s Î C in G is called an iteration group, (see e.g. [Scheinberg, 1970]), or a one-parameter group in G, if the translation equation


p(t+s,x) = p(t,p(s,x))
(T)
holds for all t,s,x Î C. Hence p(0,x) = x and p(-1,x) = p-1(1,x), the inverse of p with respect to substitution. If we express p(s,x) in the form ån ³ 1pn(s)xn, then p is called an analytic iteration group, if all the coefficient functions pn(x) are entire functions.

The formal power series p(x) is called (analytically) iterable, or embeddable, if there exists an (analytic) iteration group p in G, such that p(1,x) = p(x) for all x Î C.

There exist only three different types of analytic iteration groups in G.

  1. p(s,x) = x for all s Î C.
  2. p(s,x) = S-1(elsS(x)) for all s Î C, where l Î C\{ 0} and S(x) = x+s2x2+... belongs to G0. These iteration groups are called iteration groups of the first type. Each iteration group of this type is simultaneously conjugate to the iteration group (elsx)s Î C.
  3. p(s,x) = x+cksxk+Pk+1(k)(s)xk+1+... for all s Î C, where ck ¹ 0, k ³ 2 and Pr(k)(s) are certain polynomials in s for r > k. These iteration groups are called iteration groups of the second type.
The formal power series p(x) = x can trivially be embedded into an analytic iteration group. Assume p(x) ¹ x and p(x) = rx+c2x2+..., where r ¹ 0. If r is not a complex root of 1, then let l be a logarithm lnr. In this case there exists exactly one analytic embedding (p(s,x))s Î C of p(x), such that p(s,x) = elsx+.... Let S(x) = x+s2x2+... be the unique formal power series, such that S(p(1,S-1(x))) = rx, then p(s,x) = S-1(elsS(x)) for all s Î C.

If r is a complex root of 1 and r ¹ 1, the series p(x) need not have an analytic embedding. But if such a p(x) has an analytic embedding, then it is of the first type. In this situation, however, the embedding need not be unique.

If p(x) = x+ckxk+..., with ck ¹ 0 and k ³ 2, then there exists exactly one analytic embedding of p(x) in an iteration group of second type. (These facts about analytic iteration groups in C [[x]] can also be deduced as special cases of the results in [Reich & Schwaiger, 1977].)

Assume that a(x) and p(x) are formal power series given as above. For n Î Z we form the natural iterates of p(x) defined by


pn(x): = ì
ï
í
ï
î
x,
n = 0
p(pn-1(x)),
n > 0
(p-1)-n(x),
n < 0.
Furthermore for n ³ 0 we define


a(n,x): = n-1
Õ
r = 0 
a(pr(x))
and


b(n,x): = an(x) n-1
å
r = 0 
b(pr(x))
r
Õ
j = 0 
a(pj(x))
.
Then the conditions


a(0,x) = 1        b(0,x) = 0
(B1)


a(1,x) = a(x)        b(1,x) = b(x)
(B2)
are clearly satisfied.

Lemma 1 The two families (a(n,x))n Î N0 and (b(n,x))n Î N0 satisfy


a(n+m,x) = a(m,x)a(n,pm(x))
(C1)


b(n+m,x) = b(m,x)a(n,pm(x)) +b(n,pm(x))
(C2)
for all n,m ³ 0.

We leave the proof by induction to the reader.

If for n < 0 we define


a(n,x): = 1
a(-n,pn(x))
=


1
-n-1
Õ
r = 0 
a(pr+n(x))
= 1
-1
Õ
r = n 
a(pr(x))
and


b(n,x): = -b(-n,pn(x))
a(-n,pn(x))
= -a(n,x)b(-n,pn(x)),
then Lemma 1.1 holds for all n,m Î Z.

Lemma 2 If j(x) satisfies (L), then it also satisfies


j(pn(x)) = a(n,x)j(x)+b(n,x)
(Ln)
for all n Î Z.

Proof. Obvious from Lemma 1.1 and its generalization for all n Î Z.       [¯]

Motivated by (Ln), (C1) and (C2) for natural iterates L. Reich introduced in [Reich, 1998] the following notion.

The linear functional equation (L) has a covariant embedding with respect to the analytic iteration group (p(s,x))s Î C of p(x), if there exist families (a(s,x))s Î C and (b(s,x))s Î C of formal power series with entire coefficient functions an(s) and bn(s) for all n ³ 0, such that


j(p(s,x)) = a(s,x)j(x)+b(s,x)
(Ls)
holds for all s Î C and for all solutions j(x) of (L) in C [[x]]. Moreover it is assumed that a and b satisfy both the boundary conditions (B1) and (B2) and the cocycle equations


a(t+s,x) = a(s,x)a(t,p(s,x))
(Co1)


b(t+s,x) = b(s,x)a(t,p(s,x)) +b(t,p(s,x))
(Co2)
for all s,t Î C.

Such embeddings were studied in a much more general setting by Z. Moszner in [Moszner, 1999] and for real-valued functions by G. Guzik in [Guzik, 1999], [Guzik, 2000] and [Guzik, 2001]. For the theory of linear functional equations we refer the reader to [Kuczma et al., 1990] and to [Kuczma, 1968]. We will treat this problem in the ring of formal power series C [[x]]. In section 2 we solve the underlying functional equations (Co1) and (Co2) completely. Then in section 3 we show how to adjust these solutions to given boundary conditions. And finally, in the last section we describe how to embed the linear functional equation (L) in the generic cases.

When dealing with analytic iteration groups (p(s,x))s Î C of the first type, it is enough to consider p(s,x) = elsx. This is explained in the next

Theorem 1 Let p(s,x) = S-1(elsS(x)) for l ¹ 0 and S(x) Î G0 be an embedding of p(x).

  1. The formal power series j(x) is a solution of (L) if and only if [(j)\tilde]: = j°S-1 satisfies


    ~
    j
     
    (elsy) = ~
    a
     
    (y) ~
    j
     
    (y)+ ~
    b
     
    (y)
    ( ~
    L
     
    )
    for all s,y Î C, where [a\tilde]: = a°S-1 and [b\tilde]: = b°S-1.

  2. The system (Ls), (Co1), (Co2), (B1) and (B2) is equivalent to the system


    ~
    j
     
    (elsy) = ~
    a
     
    (s,y) ~
    j
     
    (y)+ ~
    b
     
    (s,y)
    ( ~
    L
     
    s)


    ~
    a
     
    (t+s,y) = ~
    a
     
    (s,y) ~
    a
     
    (t,elsy)
    ( ~
    C
     
    o1)


    ~
    b
     
    (t+s,y) = ~
    b
     
    (s,y) ~
    a
     
    (t,elsy) + ~
    b
     
    (t,elsy)
    ( ~
    C
     
    o2)


    ~
    a
     
    (0,y) = 1        ~
    b
     
    (0,y) = 0
    ( ~
    B
     
    1)


    ~
    a
     
    (1,y) = ~
    a
     
    (y)        ~
    b
     
    (1,y) = ~
    b
     
    (y),
    ( ~
    B
     
    2)
    where [(a)\tilde](s,y) = a(s,S-1(y)) and [(b)\tilde](s,y) = b(s,S-1(y)).

Proof. The formal series j(x) satisfies (L) if and only if


j(S-1(elsS(x))) = a(x)j(x) +b(x)Û


(j°S-1)(elsS(x)) =


(a°S-1)(S(x))(j°S-1)(S(x))+(b°S-1)(S(x)),
which is equal to ([L\tilde]) after replacing S(x) by y.

Assuming that (Ls) holds we deduce


j(S-1(elsS(x))) = a(s,x)j(x) +b(s,x)Þ


(j°S-1)(elsS(x)) =


a(s,S-1(S(x)))(j°S-1)(S(x))+b(s,S-1(S(x))),
which is equal to ([L\tilde]s) after replacing S(x) by y. The boundary conditions ([B\tilde]1) and ([B\tilde]2) are naturally equivalent to (B1) and (B2). Finally,


~
a
 
(t+s,y) = a(t+s,S-1(y)) =


a(s,S-1(y)) a(t,p(s,S-1(y))) =


~
a
 
(s,y)a(t,S-1(elsS(S-1y))) =


~
a
 
(s,y)a(t,S-1(elsy)) = ~
a
 
(s,y) ~
a
 
(t,elsy),
hence ([C\tilde]o1) is satisfied. Using similar methods it is possible to show that ([C\tilde]o2) is a consequence of (Co2).

Since (p(s,x))s Î C and (elsx)s Î C are conjugate via the formal power series S(x) it is clear how to prove the implications into the converse direction.       [¯]

2  Solutions of the cocycle equations

Lemma 3 Let E(x): = e0+e1x+... Î C [[x]], e0 ¹ 0 and let m Î C. Then


a(s,x): = ems E(p(s,x))
E(x)
is a solution of (Co1).

Proof. Since p satisfies the translation equation (T) it is clear that


a(t+s,x) = em(t+s) E(p(t+s,x))
E(x)
=


emtems E(p(t,p(s,x)))
E(x)
=


emt E(p(t,p(s,x)))
E(p(s,x))
ems E(p(s,x))
E(x)
=


a(t,p(s,x))a(s,x).
      [¯]

Lemma 2.1 also holds, when ems is replaced by a generalized exponential function.

If we express a(s,x) in the form


a(s,x) = ¥
å
n = 0 
an(s)xn,
then it follows from the cocycle equation (Co1) that a0(t+s) = a0(s)a0(t). Hence, taking into account the regularity conditions for the coefficients of a and the fact that a0(s) ¹ 0, it is clear that a0(s) = ems for some m Î C. Consequently a(s,x) = ems [^(a)](s,x) and [^(a)](s,x) = 1+[^(a)]1(s)x+... . Using the formal logarithm there exists exactly one [(a)\tilde](s,x) Î C [[x]], such that ord x [(a)\tilde](s,x) ³ 1 for all s Î C and [^(a)](s,x) = exp([(a)\tilde](s,x)). The coefficient functions of [(a)\tilde] are analytic if and only if the coefficient functions of [^(a)] are analytic, which is equivalent to the fact that the coefficient functions of a are analytic. Furthermore, [^(a)] is a solution of (Co1) if and only if [(a)\tilde] satisfies


~
a
 
(t+s,x) = ~
a
 
(s,x)+ ~
a
 
(t,p(s,x)).
(Co1¢)

Theorem 2 The family [(a)\tilde] of formal power series is a solution of (Co1¢) and [(a)\tilde](0,x) = 0 if and only if there exists a formal power series K(y) Î C [[y]], ord K(y) ³ 1, such that


~
a
 
(s,x) = ó
õ
s

0 
K(p(s,x))ds,
where integration is taken coefficientwise.

Proof. First assume that [(a)\tilde] is a solution of (Co1¢) with [(a)\tilde](0,x) = 0. Coefficientwise differentiation of (Co1¢) with respect to the variable t and the chain rule for this differentiation yields


~
a
 
¢(t+s,x) = ~
a
 
¢(t,p(s,x)).
For t = 0 we get [(a)\tilde]¢(s,x) = [(a)\tilde]¢(0,p(s,x)). Since ord x [(a)\tilde](s,x) ³ 1, also ord x [(a)\tilde]¢(s,x) ³ 1. Putting K(y): = [(a)\tilde]¢(0,y), we obtain ord K(y) ³ 1 and [(a)\tilde]¢(s,x) = K(p(s,x)). By coefficientwise integration it follows that


~
a
 
(s,x) = ó
õ
s

0 
K(p(s,x))ds.
Conversely, assume that [(a)\tilde](s,x) is given as the integral above. We prove that [(a)\tilde] satisfies (Co1¢):


~
a
 
(t+s,x) = ó
õ
t+s

0 
K(p(s,x))ds =


ó
õ
s

0 
K(p(s,x))ds+ ó
õ
t+s

s 
K(p(s,x))ds =


~
a
 
(s,x)+ ó
õ
t

0 
K(p(t+s,x))dt =


~
a
 
(s,x)+ ó
õ
t

0 
K(p(t,p(s,x)))dt =


~
a
 
(s,x)+ ~
a
 
(t,p(s,x)),
by applying (T). From the definition of [(a)\tilde] it is obvious that [(a)\tilde](0,x) = 0.       [¯]

Corollary 1 Using the notation from above, we have:

  1. The family ([^(a)](s,x))s Î C is a solution of (Co1) if and only if there exists K(y) Î C [[y]], ord K(y) ³ 1, such that


    ^
    a
     
    (s,x) = exp ó
    õ
    s

    0 
    K(p(s,x))ds.

  2. The family (a(s,x))s Î C is a solution of (Co1) if and only if there exist m Î C and K(y) Î C [[y]], ord K(y) ³ 1, such that


    a(s,x) = emsexp ó
    õ
    s

    0 
    K(p(s,x))ds.

Now we assume that a satisfies (Co1). Since ord xa(s,x) = 0, it is possible to define g(s,x) Î C [[x]] by


g(s,x): = b(s,x)
a(s,x)
       "s Î C.
The coefficient functions of g are analytic if and only if the coefficient functions of b are analytic.

Lemma 4 The families a and b satisfy the system ((Co1),(Co2)) if and only if a satisfies (Co1) and g is a solution of


g(t+s,x) = g(s,x)+ g(t,p(s,x))
a(s,x)
.
(Co2¢)

Proof. Assume first that a and b satisfy (Co1) and (Co2). Then


g(t+s,x) = b(t+s,x)
a(t+s,x)
=


b(s,x)a(t,p(s,x)) +b(t,p(s,x))
a(s,x)a(t,p(s,x))
=


b(s,x)
a(s,x)
+ b(t,p(s,x))
a(s,x)a(t,p(s,x))
=


g(s,x)+ g(t,p(s,x))
a(s,x)
.
Assuming conversely that g is a solution of (Co2¢) we get


b(t+s,x) = a(t+s,x)g(t+s,x) =


a(s,x)a(t,p(s,x))[g(s,x)+ g(t,p(s,x))
a(s,x)
] =


a(s,x)g(s,x)a(t,p(s,x))+


a(t,p(s,x))g(t,p(s,x)) =


b(s,x)a(t,p(s,x))+ b(t,p(s,x)) .
      [¯]

Theorem 3 Assume that a satisfies the cocycle equation (Co1). Then a and b are a solution of (Co2) if and only if there exists a series L(y) Î C [[y]], such that


b(s,x) = a(s,x) ó
õ
s

0 
L(p(s,x))
a(s,x)
ds,
where integration is taken coefficientwise.

Proof. First assume that a and b satisfy (Co2). Then Lemma 2.4 implies that a and g satisfy (Co2¢). Coefficientwise differentiation of (Co2¢) with respect to the variable t yields


g¢(t+s,x) = g¢(t,p(s,x))
a(s,x)
.
For t = 0 we get g¢(s,x) = g¢(0,p(s,x))/a(s,x). Putting L(y): = g¢(0,y), we obtain g¢(s,x) = L(p(s,x))/a(s,x). By coefficientwise integration it follows that


g(s,x) = ó
õ
s

0 
L(p(s,x))
a(s,x)
ds
and


b(s,x) = a(s,x) ó
õ
s

0 
L(p(s,x))
a(s,x)
ds.
Conversely, if b is given by that formula, then


g(t+s,x) = ó
õ
t+s

0 
L(p(s,x))
a(s,x)
ds =


ó
õ
s

0 
L(p(s,x))
a(s,x)
ds+ ó
õ
t+s

s 
L(p(s,x))
a(s,x)
ds =


g(s,x)+ ó
õ
t

0 
L(p(t+s,x))
a(t+s,x)
dt =


g(s,x)+ ó
õ
t

0 
L(p(t,p(s,x)))
a(s,x)a(t,p(s,x))
dt =


g(s,x)+ g(t,p(s,x))
a(s,x)
.
In other words, a and g satisfy (Co2¢), hence by Lemma 2.4 a and b satisfy (Co2).       [¯]

Now we will describe a different form of representing the general solution of (Co1), and of the system ((Co1),(Co2)), involving as few integrals as possible. In Lemma 2.1 we already derived solutions a of (Co1) which could be represented without integrals at all. Their form is a motivation for the representation of the general solution of (Co1) we have in mind here. In the first part of Lemma 2.7 we will, similarly, present a class of solutions of ((Co1),(Co2)), which are free of integrals. This motivates the representation of the general solution of ((Co1),(Co2)) and will be applied in the proof of the form of the general solution. In this context it will be necessary and helpful to distinguish between the different types of iteration groups (p(s,x))s Î C, and also to consider certain special cases of m and of (m,l), if iteration groups of the first type are used. In particular, we will investigate under which conditions the solutions can be expressed without integrals. Theorem 2.6 summarizes our results concerning (Co1), Theorem 2.8 the results concerning the system ((Co1),(Co2)). The above mentioned form of the general solutions will be useful in solving the boundary conditions.

Theorem 4

  1. Let p(s,x) = elsx for l ¹ 0. Then a is a solution of (Co1) if and only if there exist m Î C and a formal power series E(x) = 1+e1x+... Î C [[x]], such that


    a(s,x) = ems E(elsx)
    E(x)
    .
    The series E(x) is uniquely determined by a.

  2. Let p(s,x) = x+cksxk+... Î C [[x]] with ck ¹ 0 and k ³ 2. If a(s,x) = ems(1+[^(a)]k(s)xk+...) º ems mod ord x k, then a is a solution of (Co1) if and only if there exist m Î C and a series E(x) = 1+e1x+... Î C [[x]], such that


    a(s,x) = ems E(p(s,x))
    E(x)
    .
    The series E(x) is uniquely determined by a.

  3. The general solution a of (Co1) for iteration groups (p(s,x))s Î C of second type is


    a(s,x) =


    ems k-1
    Õ
    n = 1 
    æ
    è
    exp ó
    õ
    s

    0 
    p(s,x)nds ö
    ø
    kn
     
    E(p(s,x))
    E(x)
    ,
    with arbitrary kn Î C.

Proof. In Lemma 2.1 we described solutions a of (Co1) which could be expressed without integrals. In Corollary 2.3 all solutions of this equation in integral form were determined. Combining these two results we investigate when


emsexp ó
õ
s

0 
K(p(s,x))ds = ems E(p(s,x))
E(x)
(1)
holds, where E(x) º 1 mod ord 1. After applying the formal logarithm we have to check when


ó
õ
s

0 
K(p(s,x))ds = ~
E
 
(p(s,x))- ~
E
 
(x)
is true, for [E\tilde](x): = lnE(x). Coefficientwise differentiation of the last equation with respect to the variable s yields


K(p(s,x)) =
d ~
E
 

dy
ê
ê

y = p(s,x) 
p¢(s,x),
(2)
where we used the ``mixed'' chain rule for this derivation. If (p(s,x))s Î C is an iteration group of the first type, this means


K(elsx) =
d ~
E
 

dy
ê
ê

y = elsx 
lelsx .
In this formula elsx can be replaced by the indeterminate y, hence we get K(y) = ly[d[E\tilde](y)/dy]. Since ord K(y) ³ 1 and l ¹ 0, it is possible to divide by ly and we end up with a differential equation


K(y)
ly
= : ~
K
 
(y) =
d ~
E
 
(y)

dy
.
(3)
Assume that [K\tilde](y) = ån ³ 0 [(k)\tilde]nyn, then the Ansatz [E\tilde](y) = ån ³ 1 [e\tilde]nyn leads to [e\tilde]n+1 = [(k)\tilde]n/(n+1) for all n ³ 0. Hence, all the coefficients en of E(x) for n ³ 1 and n = 0 are uniquely determined.

So far we proved that the series E is uniquely defined by K. Next we show that each solution [E\tilde] of the differential equation (3) with [E\tilde](0) = 0 leads to a solution E of (1) by setting E(y) = exp[E\tilde](y). Since [E\tilde] is a solution of the differential equation, it is clear that


K(elsx) = lelsx
d ~
E
 

dy
ê
ê

y = elsx 
.
The right hand side of this equation is [()/(s)][E\tilde](elsx), hence


ó
õ
s

0 
K(elsx) ds = ~
E
 
(elsx) - ~
E
 
(el0 x)
and


emsexp ó
õ
s

0 
K(elsx) ds =


emsexp( ~
E
 
(elsx) - ~
E
 
(x)) =


ems
exp ~
E
 
(elsx)

exp ~
E
 
(x)
= ems E(elsx)
E(x)
.
Moreover, the coefficient e0 of E(x) is equal to 1, since E(x) = exp[E\tilde](X) and ord [E\tilde](x) ³ 1, which finishes the proof for iteration groups p of the first type.

If (p(s,x))s Î C is an analytic iteration group of the second type, then from iteration theory (cf. [Scheinberg, 1970] or [Reich & Schwaiger, 1977]) it follows that p¢(s,x) = H(p(s,x)), where H(y): = p¢(s,y)|s = 0 is the infinitesimal generator of p. In the present situation H(y) = ckyk+..., hence ord H(y) = k, and (2) means


K(p(s,x)) =
d ~
E
 

dy
ê
ê

y = p(s,x) 
H(p(s,x)).
After replacing p(s,x) by the indeterminate y we realize that ord K(y) ³ k, since K(y) = H(y) [d[E\tilde](y)/dy]. (This however is equivalent to a(s,x) º ems mod ord x k.) Hence we end up with the differential equation


K(y)
H(y)
= : ~
K
 
(y) =
d ~
E
 
(y)

dy
,
(4)
which, similar as in the first part of the proof, has exactly one solution [E\tilde](y) = ån ³ 1 [e\tilde]nyn.

Finally it remains to prove that each solution [E\tilde] of this differential equation with [E\tilde](0) = 0 yields a solution E of (1). Let [E\tilde] be a solution of (4) with [E\tilde](0) = 0, then


K(p(s,x)) =
d ~
E
 

dy
ê
ê

y = p(s,x) 
H(p(s,x)) =


d ~
E
 

dy
ê
ê

y = p(s,x) 
p¢(s,x) =
s
~
E
 
(p(s,x))
and


ó
õ
s

0 
K(p(s,x)) ds = ~
E
 
(p(s,x))- ~
E
 
(p(0,x)) =


~
E
 
(p(s,x))- ~
E
 
(x).
Substitution into the exponential series and multiplication by ems yields


emsexp ó
õ
s

0 
K(p(s,x)) ds =


emsexp( ~
E
 
(p(s,x))- ~
E
 
(x)) = ems E(p(s,x))
E(x)
.
Hence E(x) satisfies (1) and E(x) = exp[E\tilde](x) º 1 mod ord 1.

From Corollary 2.3 we deduce that the general solution a of (Co1) is given by


a(s,x) =


emsexp ó
õ
s

0 
æ
è
k-1
å
n = 1 
knp(s,x)n+
å
n ³ k 
knp(s,x)n ö
ø
ds =


emsexp æ
è
k-1
å
n = 1 
kn ó
õ
s

0 
p(s,x)nds ö
ø
·


·exp ó
õ
s

0 
^
K
 
(p(s,x))ds
with [^K](y) = ån ³ kknyn. By Corollary 2.3


exp ó
õ
s

0 
^
K
 
(p(s,x))ds
is a solution of (Co1), and it is of the form 1+[^(a)]k(s)xk+..., since ord [^K](y) ³ k. Hence, by the second part of the present theorem there exists a unique series E(x) = 1+e1x+..., such that


exp ó
õ
s

0 
^
K
 
(p(s,x))ds = E(p(s,x))
E(x)
.
Summarizing, we found


a(s,x) =


ems k-1
Õ
n = 1 
æ
è
exp ó
õ
s

0 
p(s,x)nds ö
ø
kn
 
E(p(s,x))
E(x)
.
      [¯]

Lemma 5 Let E(x): = e0+e1x+... Î C [[x]], e0 ¹ 0, and assume that F(x) Î C [[x]] and m Î C.

  1. The series


    b(s,x): =


    emsE(p(s,x))[F(x)-e-msF(p(s,x))]
    together with a given in Lemma 2.1 satisfies (Co2) for any analytic iteration group p.

  2. Assume that p(s,x) = x+cksxk+... Î C [[x]] with k ³ 2 and ck ¹ 0 is an analytic iteration group of the second type, and let P(s,x) denote the series


    P(s,x): = k-1
    Õ
    n = 1 
    æ
    è
    exp ó
    õ
    s

    0 
    p(s,x)nds ö
    ø
    kn
     
    .
    Then b defined by


    b(s,x): =


    emsP(s,x)E(p(s,x))[F(x)-e-ms F(p(s,x))
    P(s,x)
    ]
    together with a given in the third part of Theorem 2.6 satisfies (Co2).

Proof. The families a and b satisfy (Co2) if and only if


b(t+s,x)-b(t,p(s,x)) = b(s,x)a(t,p(s,x))
for all s,t Î C. If we express b and a by E, F, p, this is


em(t+s) E(p(t+s,x))[F(x)-e-m(t+s)F(p(t+s,x))]-


emt E(p(t,p(s,x))) é
ë
F(p(s,x))-


e-mtF(p(t,p(s,x))) ù
û
=


emsE(p(s,x))[F(x)-e-msF(p(s,x))]


emt E(p(t,p(s,x)))
E(p(s,x))
.
Application of (T) together with simplification of both sides yields


em(t+s) E(p(t+s,x))F(x)


-emt E(p(t,p(s,x)))F(p(s,x)) =


emsF(x)emtE(p(t,p(s,x)))-


F(p(s,x))emt E(p(t,p(s,x))),
which is always true since p satisfies (T).

The proof of the second part is similar to the proof above, the reader only has to take into account that (P(s,x))s Î C is a solution of (Co1).       [¯]

Theorem 5 Let a be a solution of (Co1).

  1. Assume that p(s,x) = elsx for l ¹ 0 and that a is given as in the first part of Theorem 2.6.

    If m-nl ¹ 0 for all n Î N0, then (a,b) is a solution of (Co2) if and only if there exists a formal power series F(x) Î C [[x]], such that


    b(s,x) = emsE(elsx)[F(x)-e-msF(elsx)].
    The series F(x) is uniquely determined by a and b.

    If m = n0l for n0 Î N0, then (a,b) is a solution of (Co2) if and only if there exist a formal power series F(x) Î C [[x]] and ln0 Î C, such that


    b(s,x) =


    emsE(elsx)[ln0sxn0+F(x)-e-msF(elsx)].

  2. Let p(s,x) = x+cksxk+... Î C [[x]] with ck ¹ 0 and k ³ 2.

    Assume that a can be expressed as in the second part of Theorem 2.6. If m ¹ 0, then (a,b) is a solution of (Co2) if and only if there exists a formal power series F(x) Î C [[x]], such that


    b(s,x) =


    emsE(p(s,x))[F(x)-e-msF(p(s,x))].
    The series F(x) is uniquely determined by a and b.

    If m = 0, then (a,b) is a solution of (Co2) if and only if there exists a series F(x) Î C [[x]], such that


    b(s,x) = E(p(s,x))[F(x)-F(p(s,x))].
    Furthermore ord xb(s,x) ³ k.

    Let a be the general solution of (Co1) given in the third part of Theorem 2.6, and assume that


    P(s,x): = k-1
    Õ
    n = 1 
    æ
    è
    exp ó
    õ
    s

    0 
    p(s,x)nds ö
    ø
    kn
     
    actually occurs. Then let n0 be the minimum of { n | 1 £ n £ k-1,  kn ¹ 0} . If m ¹ 0 or n0 ¹ k-1 or kk-1-m ck ¹ 0 for all m Î N, then (a,b) is a solution of (Co2) if and only if there exists a formal power series F(x) Î C [[x]], such that


    b(s,x) =


    emsP(s,x)E(p(s,x))[F(x)-e-ms F(p(s,x))
    P(s,x)
    ].
    The series F(x) is uniquely determined by a and b, and furthermore ord xb(s,x) ³ n0.

    If m = 0, n0 = k-1 and kk-1 = n1 ck for n1 Î N, then (a,b) is a solution of (Co2) if and only if there exists a series F(x) Î C [[x]] and ln1+n0¢¢ Î C, such that b(s,x) equals


    E(p(s,x))P(s,x)·[ F(x)-


    F(p(s,x))
    P(s,x)
    +ln1+n0¢¢ ó
    õ
    s

    0 
    p(s,x)n1+n0
    P(s,x)
    ds ù
    ú
    û
    .
    Also in this situation ord xb(s,x) ³ k-1.

Proof. We apply similar ideas and arguments as in the proof of Theorem 2.6. In Lemma 2.7 we described special solutions, in Theorem 2.5 all solutions of (Co2) in integral form. If a can be expressed without any integrals, then we check when


a(s,x) ó
õ
s

0 
L(p(s,x))
a(s,x)
ds =
emsE(p(s,x))[F(x)-e-msF(p(s,x))]
(5)
holds. This is equivalent to


ó
õ
s

0 
L(p(s,x))
a(s,x)
ds = E(x)[F(x)-e-msF(p(s,x))].
Coefficientwise differentiation of the last equation with respect to the variable s yields


L(p(s ,x))
a(s,x)
=


E(x)[me-msF(p(s,x))-e-ms d F
d y
ê
ê

y = p(s,x) 
p¢(s,x)].
If p is an iteration group of the first type this means


L(elsx) =


a(s,x)E(x)[me-msF(elsx)-e-ms d F
d y
ê
ê

y = elsx 
lelsx].
Using the special form of a from the first part of Theorem 2.6 and replacing elsx by the indeterminate y gives


~
L
 
(y): = L(y)
E(y)
= mF(y)-ly dF(y)
dy
.
(6)
Assume that [L\tilde](y) = ån ³ 0 lnyn, then the Ansatz F(y) = ån ³ 0 fnyn leads to



å
n ³ 0 
lnyn =
å
n ³ 0 
(m-nl) fnyn.
If m-nl ¹ 0 for all n ³ 0, then F(y) is uniquely given by


fn = ln
m-nl
       "n ³ 0.

So far we proved that in this situation the series F is uniquely defined by L. Next we show that each solution F of the differential equation (6) is a solution of (5). Since F is a solution of (6) it is clear that


L(y) = E(y)[mF(y)-ly dF(y)
dy
].
After replacing y by elsx and using the special form of a we derive that


L(elsx)
a(s,x)
=


E(x)[me-msF(elsx)-e-mslelsx dF
dy
ê
ê

y = elsx 
] =


E(x)
s
(-e-msF(elsx)).
Coefficientwise integration finally yields the desired result.

Still we are dealing with analytic iteration groups (p(s,x))s Î C of the first type. But now we assume that m = n0l. In this situation comparing the coefficients of yn0 yields the condition 0 = (m-n0l)fn0 = ln0. If ln0 ¹ 0, then we split L(y) in the form



å
[(n ³ 0) || (n ¹ n0)] 
ln yn +ln0 yn0.
From Theorem 2.6 we know that b is given as


b(s,x) = a(s,x) ó
õ
s

0 
L(elsx)
a(elsx)
ds =


ems E(elsx)
E(x)
ó
õ
s

0 
L(elsx)E(x)
emsE(elsx)
ds =


emsE(elsx) ó
õ
s

0 
e-ms
å
n ³ 0 
lnenlsxnds =


emsE(elsx) æ
è
ó
õ
s

0 
ln0ds  xn0 +


+ ó
õ
s

0 

å
[(n ³ 0) || (n ¹ n0)] 
e(nl-m)s ln xn ds ö
ø
=


emsE(elsx)[ln0sxn0+F(x)-e-msF(elsx)].
For n ¹ n0 the coefficients fn of F(x) are uniquely given by


fn = ln
m-nl
,
whereas fn0 can be arbitrarily chosen in C.

In the next part of the proof we assume that (p(s,x))s Î C is an iteration group of the second type. We investigate when (5) holds. For doing this we assume that a is given as in the second part of Theorem 2.6. Inserting the special form of a, coefficientwise differentiation with respect to s and expressing p¢(s,x) as H(p(s,x)), where H is the infinitesimal generator of p, yields the equation


L(p(s,x)) =


E(p(s,x))[mF(p(s,x))- dF
dy
ê
ê

y = p(s,x) 
H(p(s,x))].
After replacing p(s,x) by y we end up with the differential equation


~
L
 
(y): = L(y)
E(y)
= mF(y)- dF(y)
dy
H(y).
(7)
Assume that [L\tilde](y) = ån ³ 0 lnyn and H(y) = ån ³ khnyn, where hk = ck ¹ 0, then the Ansatz F(y) = ån ³ 0 fnyn leads to



å
n ³ 0 
lnyn =


m k-1
å
n = 0 
fn yn+
å
n ³ k 
æ
è
mfn -
å
[(r+s = n) || (r ³ k)] 
(r+1)fr+1hs ö
ø
yn.
Comparing coefficients yields


ln
=
mfn,
n < k
ln
=
mfn - n-k+1
å
r = 1 
r fr hn-r+1,   
n ³ k.
If m ¹ 0, then F is uniquely determined by


fn
=
ln
m
,
n < k
fn
=
1
m
æ
è
ln+ n-k+1
å
r = 1 
r fr hn-r+1 ö
ø
,    
n ³ k.
Assuming conversely that F is a solution of (7), then it is left to the reader to prove that F satisfies (5). (The proof is similar to that given in the first part of this proof.)

If m = 0, then (7) reduces to


~
L
 
(y) = - dF(y)
dy
H(y)
(8)
or in more details



å
n ³ 0 
lnyn = -
å
n ³ k 
æ
è
n-k+1
å
r = 1 
r fr hn-r+1 ö
ø
yn.
Comparing coefficients yields a necessary condition for the coefficients of [L\tilde], namely


ln = 0,        n < k
and a formula to determine recursively the values of fn by


fn = -
lk+n-1+ n-1
å
r = 1 
r fr hk+n-r

n ck
,       n ³ 1,
since hk = ck. The value f0 can be arbitrarily chosen in C. Hence ord [L\tilde](y) ³ k, which implies that ord L(y) ³ k and finally ord x b(s,x) ³ k.

Assuming conversely that F is a solution of (8), then it is left to the reader to prove that F satisfies (5) for m = 0.

Finally, let a be the general solution of (Co1). Then


b(s,x) =


emsP(s,x) E(p(s,x))
E(x)
ó
õ
s

0 
L(p(s,x))E(x)
emsP(s,x)E(p(s,x))
ds =


emsP(s,x)E(p(s,x)) ó
õ
s

0 
e-ms
~
L
 
(p(s,x))

P(s,x)
ds.
Now we have to check when


emsP(s,x)E(p(s,x)) ó
õ
s

0 
e-ms
~
L
 
(p(s,x))

P(s,x)
ds =
emsP(s,x)E(p(s,x))[F(x)-e-ms F(p(s,x))
P(s,x)
]
(9)
holds. This is obviously equivalent to


ó
õ
s

0 
e-ms
~
L
 
(p(s,x))

P(s,x)
ds = F(x)-e-ms F(p(s,x))
P(s,x)
.
Coefficientwise differentiation with respect to the variable s gives


e-ms
~
L
 
(p(s,x))

P(s,x)
= -
s
e-ms F(p(s,x))
P(s,x)
=


me-ms F(p(s,x))
P(s,x)
-e-ms æ
ç
è

s
1
P(s,x)
ö
÷
ø
F(p(s,x))-


e-ms 1
P(s,x)
dF
dy
ê
ê

y = p(s,x) 
H(p(s,x)),
where H is the infinitesimal generator of p. Since



s
1
P(s,x)
=
k-1
å
n = 1 
(-kn)p(s,x)n

P(s,x)
we end up with the following differential equation after replacing p(s,x) by y,


~
L
 
(y) = æ
è
m+ k-1
å
n = 1 
kn yn ö
ø
F(y)- dF(y)
dy
H(y).
(10)
The usual Ansatz leads to



å
n ³ 0 
ln = m
å
n ³ 0 
fn yn+
å
n ³ 1 
æ
è
min{ k-1,n}
å
r = 1 
kr fn-r ö
ø
yn-



å
n ³ k 
æ
è
n-k+1
å
r = 1 
r fr hn-r+1 ö
ø
yn.
Hence the coefficients satisfy


l0 = mf0,


ln = mfn+ n
å
r = 1 
kr fn-r
for 1 £ n < k, and


ln = mfn+ k-1
å
r = 1 
kr fn-r - n-k+1
å
r = 1 
r fr hn-r+1
for n ³ k. If m ¹ 0, then F is uniquely given by


f0 = l0
m
,


fn = 1
m
æ
è
ln - n
å
r = 1 
kr fn-r ö
ø
for 1 £ n < k , and


fn = 1
m
æ
è
ln - k-1
å
r = 1 
kr fn-r + n-k+1
å
r = 1 
r fr hn-r+1 ö
ø
for n ³ k.

Again it is left to the reader to prove that each of these solutions F satisfies (9).

What happens in the case m = 0? If n0 denotes min{ n | 1 £ n £ k-1,  kn ¹ 0} , then (10) reduces to


~
L
 
(y) = æ
è
k-1
å
n = n0 
kn yn ö
ø
F(y)- dF(y)
dy
H(y).
Since the right hand side is a power series of order ³ n0, the coefficients of [L\tilde](y) satisfy


ln = 0
for 0 £ n < n0,


ln = n
å
r = n0 
kr fn-r
for n0 £ n < k and


ln = k-1
å
r = n0 
kr fn-r - n-k+1
å
r = 1 
r fr hn-r+1
for n ³ k.

Consequently ord [L\tilde](y) ³ n0 and ord xb(s,x) ³ n0. Hence for n0 £ n < k the coefficients fn-n0 are uniquely determined by


fn-n0 = 1
kn0
æ
è
ln- n
å
r = n0+1 
kr fn-r ö
ø
.
(11)
For n ³ k we still have to consider different cases. If n0 < k-1, then n-n0 > n-k+1 and fn-n0 are uniquely given by the recursive formula


fn-n0 =


1
kn0
æ
è
ln- k-1
å
r = n0+1 
kr fn-r+ n-k+1
å
r = 1 
r fr hn-r+1 ö
ø
.

If n0 = k-1, then for n ³ k


ln = kk-1fn-k+1- n-k+1
å
r = 1 
r frhn-r+1 =
(kk-1- (n-k+1)hk)fn-k+1- n-k
å
r = 1 
r fr hn-r+1.
(12)
Hence, if kk-1-mck ¹ 0 for all m Î N, then


fn-n0 = fn-k+1 =
ln + n-k
å
r = 1 
r fr hn-r+1

kk-1- (n-k+1)ck
(13)
for n ³ k.

Finally we have to consider the case that m = 0, n0 = k-1 and there exists n1 Î N, such that kk-1 = n1ck. Then (12) means


ln = (kk-1-(n-k+1)ck) fn-k+1 - n-k
å
r = 1 
r fr hn-r+1
for n ³ k. If n-k+1 = n1, which is equivalent to n = n1+k-1, we have


ln1+k-1 = (kk-1-n1ck) fn1 - n1-1
å
r = 1 
r fr hn1+k-r =


- n1-1
å
r = 1 
r fr hn1+k-r.
This is a necessary condition for writing b(s,x) in the above form. In general let


ln1+k-1¢: = - n1-1
å
r = 1 
r fr hn1+k-r,
then


ó
õ
s

0 
~
L
 
(p(s,x))

P(s,x)
ds = ó
õ
s

0 
1
P(s,x)

å
n ³ k 
lnp(s,x)nds =


ó
õ
s

0 
æ
ç
ç
ç
è

å
[(n ³ k) || (n ¹ n1+k-1)] 
lnp(s,x)n+ln1+k-1¢p(s,x)n1+k-1

P(s,x)
+


(ln1+k-1-ln1+k-1¢)p(s,x)n1+k-1
P(s,x)
ö
÷
ø
ds =


F(x)- F(p(s,x))
P(s,x)
+ln1+k-1¢¢ ó
õ
s

0 
p(s,x)n1+k-1
P(s,x)
ds,
where ln1+k-1¢¢ = ln1+k-1-ln1+k-1¢. For n ¹ n1+k-1 the coefficients fn-k+1 of the series F(x) are uniquely given by the two formulae (11) and (13), and fn1 can be arbitrarily chosen in C.

Also in the last cases it is left to the reader to prove that each solution F of the differential equation also satisfies (9) for m = 0.       [¯]

3  Solutions which satisfy the boundary conditions

In this section we assume that (p(s,x))s Î C is a given iteration group. We want to determine solutions a and b of the cocycle equations (Co1) and (Co2) which also satisfy the boundary conditions (B1) and (B2) for given formal power series


a(x) =
å
n ³ 0 
anxn,  a0 ¹ 0 and b(x) =
å
n ³ 0 
bnxn.
From the results of the previous section it is obvious that (B1) is always satisfied. We only have to consider (B2) for further investigations.

First we will deal with analytic iteration groups of the first type, i.e. we consider p(s,x) = elsx for l ¹ 0. Before describing the solutions a which satisfy the boundary conditions we need a preliminary result. If J = J(l) denotes the set { n Î N | nl Î 2Zpi} , then the following lemma holds.

Lemma 6 Assume that J is not empty and let j0 be the minimum of J. Then J = Nj0.

Proof. Since j0 Î J, there exists z0 Î Z, such that j0l = 2z0pi. Then nj0l = 2nz0pi Î 2Zpi for all n Î N. Hence Nj0 is a subset of J. Conversely, assume that n Î J, then by division we deduce that n = qj0+r with uniquely determined r, such that 0 £ r < j0. From 2Zpi ' nl = (qj0+r)l = qj0l+rl = 2qz0pi+rl it follows that rl Î 2Zpi and consequently r = 0. Hence n Î Nj0, which finishes the proof.       [¯]

In the first part of Theorem 2.6 the general solution a of (Co1) for analytic iteration groups (p(s,x))s Î C of the first type was described. We want to analyze how to adjust it to the condition a(1,x) = a(x).

Theorem 6 Assume that a(x) is a given formal power series of order 0.

If J = Æ, then there exists exactly one formal power series E(x), such that


a(s,x) = ems E(elsx)
E(x)
is a solution of (Co1) and satisfies a(1,x) = a(x).

If J ¹ Æ, then there exist formal power series E(x), such that a(s,x) of the above form satisfies both (Co1) and the boundary condition if and only if for n Î J the coefficients an satisfy


an = - n-1
å
r = 1 
aren-r,
where en are the coefficients of E(x).

Proof. Writing a as indicated above, the assumption a(1,x) = a(x) is equivalent to


em
å
n ³ 0 
enenl xn =
å
n ³ 0 
n
å
r = 0 
aren-r.
Comparing coefficients yields for n = 0 that em = a0, since e0 = 1. Hence m is a logarithm lna0. For n ³ 1 we get


emenlen = a0en+ n-1
å
r = 1 
aren-r+an,
which implies


a0(enl -1)en = n-1
å
r = 1 
aren-r+an.
For n Ï J the coefficient en is uniquely determined by


en =
n
å
r = 1 
aren-r

a0(enl-1)
.
However, for n Î J the coefficient en can be chosen arbitrarily in C and an must satisfy the condition above. We will not analyze these conditions further in this paper.       [¯]

Before we adjust b to the condition b(1,x) = b(x) we need another preliminary result. Let K = K(m,l) denote the set { n Î N0 | m-nl Î 2Zpi} . Then the following lemma holds.

Lemma 7 Assume that the cardinality of K is greater than 1. Then J is not empty and


K = { k0+nj0 | n Î N} =


{ n Î N | n º k0 mod j0} ,
where k0: = minK and j0: = minJ. If | K| = 1, then J = Æ.

Proof. First we prove that when n1 and n2 are two different elements of K such that n1 > n2, then n1-n2 belongs to J. Since n1,n2 Î K, there exist z1,z2 Î Z, such that m-n1l = 2z1pi and m-n2l = 2z2pi. Then (n1-n2)l = (m-n2l)-(m-n1l) = 2(z2-z1)pi Î 2Zpi and n1-n2 Î N. Hence n1-n2 Î J.

Since k0 Î K and j0 Î J, there exist z0,z1 Î Z, such that m-k0l = 2z0pi and j0l = 2z1pi. Let n Î N0, then m-(k0+nj0)l = m-k0l-nj0l = 2z0pi-2nz1pi = 2(z0-nz1)pi Î 2Zpi and consequently k0+nj0 Î K. Thus k0+N0j0 Í K.

In the next step we prove that k0 < j0. (Then it is clear that k0+N0j0 is the set of all positive integers congruent k0 modulo j0.) If we assume that k0-j0 ³ 0, then k0-j0 Î K since m- (k0-j0)l = 2(z0+z1)pi Î 2Zpi. Moreover k0-j0 < k0 which is a contradiction to the construction of k0.

Finally we have to prove that K Í k0+N0j0. Let n Î K. If n ¹ k0, then n > k0 and then there exists z Î Z, such that m-nl = 2zpi. Moreover (n-k0)l = 2(z0-z)pi Î 2Zpi, thus n-k0 Î J = Nj0 by Lemma 3.1. Hence n Î k0+Nj0.

If | K| = 1, then J = Æ. If we assume that J ¹ Æ and k0 Î K, then J = Nj0. Hence k0+J Ì K, which is a contradiction to | K| = 1.       [¯]

Let a be a solution of (Co1) where p is an analytic iteration group of the first type. The general form of b, which satisfies together with a the cocycle equation (Co2), was given in the first part of Theorem 2.8.

Theorem 7 Let p be an analytic iteration group of the first type. Assume that b(x) is a given formal power series and a is a solution of (Co1) given by


a(s,x) = ems E(elsx)
E(x)
.

  1. If K = Æ, then there exists exactly one formal power series F(x), such that


    b(s,x) = emsE(elsx)[F(x)-e-msF(elsx)]
    together with a is a solution of (Co2) satisfying b(1,x) = b(x).

  2. If K ¹ Æ and m-nl ¹ 0 for all n Î N0, then there exist formal power series F(x), such that


    b(s,x) = emsE(elsx)[F(x)-e-msF(elsx)]
    together with a is a solution of (Co2) satisfying the boundary condition if and only if for n Î K the coefficients bn satisfy


    bn = n-1
    å
    r = 0 
    en-rfre(n-r)l(em-erl),
    (14)
    where en and fn are the coefficients of E(x) and F(x).

  3. If K ¹ Æ and m-n0l = 0, then there exist formal power series F(x) and ln0 Î C, such that


    b(s,x) =


    emsE(elsx)[ln0sxn0+F(x)-e-msF(elsx)]
    together with a is a solution of (Co2) satisfying the boundary condition if and only if for n Î K\{ n0} the coefficients bn satisfy (14) for n < n0, and bn equals


    n-1
    å
    [r = 0 || (r ¹ n0)] 
    en-rfre(n-r)l(em-erl) +ln0en-n0enl
    for n > n0.

Proof. Writing b in the form


b(s,x) = emsE(elsx)[F(x)-e-msF(elsx)],
and assuming that there is no n0 Î N such that m = n0l, then the assumption b(x) = b(1,x) is equivalent to



å
n ³ 0 
bnxn = em
å
n ³ 0 
æ
è
n
å
r = 0 
en-rfre(n-r)l ö
ø
xn-



å
n ³ 0 
æ
è
n
å
r = 0 
en-rfr ö
ø
enlxn,

which yields for all n ³ 0 that bn equals


n-1
å
r = 0 
en-rfre(n-r)l(em-erl)+e0fnenl(em-nl-1).
If n Ï K, then fn is uniquely given by


fn =
bn- n-1
å
r = 0 
en-rfre(n-r)l(em-erl)

enl(em-nl-1)
.
For n Î K the coefficient fn can be arbitrarily chosen in C and bn must satisfy


bn = n-1
å
r = 0 
en-rfre(n-r)l(em-erl).

If there is n0 Î N, such that m = n0l, then n0 Î K. Since in this situation


b(s,x) = emsE(elsx)[ln0sxn0+F(x)-e-msF(elsx)],
the formulae above are only correct for n < n0. Comparing coefficients for n ³ n0 yields


bn = n
å
[r = 0 || (r ¹ n0)] 
en-rfre(n-r)l(em-erl) +ln0en-n0enl.
For n = n0 the coefficient ln0 is uniquely determined by


ln0 = e-m æ
è
bn0- n0-1
å
r = 0 
en0-rfre(n0-r)l(em-erl) ö
ø
.
Furthermore, for n > n0 and n\not Î K, the coefficient fn is given by the fraction


bn- n-1
å
[r = 0 || (r ¹ n0)] 
en-rfre(n-r)l(em-erl)-ln0en-n0enl

enl(em-nl-1)
and for n > n0, n Î K the bn must satisfy the condition


bn = n-1
å
[r = 0 || (r ¹ n0)] 
en-rfre(n-r)l(em-erl) +ln0en-n0enl.
We will not analyze these conditions further.       [¯]

Theorem 8 Let p(s,x) = els x, where r = el is not a complex root of 1. Assume that a(x) and b(x) are given power series, where ord a(x) = 0. For each a, which satisfies (Co1) and the two boundary conditions, there exists exactly one b, such that the pair (a,b) is a solution of (Co1) and (Co2), which also satisfies the boundary conditions (B1) and (B2).

Proof. Since r is not a complex root of 1 it is obvious that J = J(l) is empty. Hence, according to Lemma 3.3 the set K = K(m,l) is empty or K has cardinality 1. In the first case everything is clear from Theorem 3.4. If K = { k0} , there exists some z Î Z, such that m-k0l = 2zpi. Then m-2zpi = k0l. If we replace m by m¢: = m-2zpi, then a0 = em = em¢ and m¢ = k0l, which means that K(m¢,l) = { k0} and m¢-k0l = 0, hence k0 = n0 from the second part of Theorem 3.4. As was described in the proof of Theorem 3.4 the coefficients fn (for n ¹ k0) and lk0 can uniquely be determined. Just fk0 can be arbitrarily chosen in C. Moreover the series b(x) need not satisfy any necessary conditions, so also in this situation there always exists a b satisfying (Co2) and (B2). According to Theorem 2.8 for computing b(s,x) it is necessary to determine F(x)-e-m¢sF(elsx). Because of the special choice of m¢ and l this difference reads as



å
n ³ 0 
fnxn-e-m¢s
å
n ³ 0 
fnenlsxn =



å
n ³ 0 
(1-e(-m¢+nl)s)fnxn =
å
[(n ³ 0) || (n ¹ k0)] 
(1-e(-m¢+nl)s)fnxn,
consequently it does not depend on the coefficient fk0, which still could be chosen arbitrarily. Hence b is uniquely determined in this situation.       [¯]

Now we come back to the analytic iteration groups of second type, i.e. p(s,x) = x+cksxk+... with k ³ 2 and ck ¹ 0. The embedding for those a, which are of the form a(s,x) = ems+ak(s)xk+..., is described in

Theorem 9 Assume that a(x) is a given formal power series of order 0, and p is an analytic iteration group of the second type. There exists exactly one formal power series E(x) = 1+e1x+..., such that


a(s,x) = ems E(p(s,x))
E(x)
is a solution of (Co1) satisfying the boundary condition if and only if an = 0 for 1 £ n < k.

Proof. In this situation again the boundary condition a(1,x) = a(x) is equivalent to a(x)E(x) = emE(p(x)). First we compute E(p(x)) which is equal to



å
n ³ 0 
en(p(x))n =
å
n ³ 0 
en(x+ckxk+...)n =



å
n ³ 0 
en(xn+ æ
ç
è
n
1
ö
÷
ø
ck xn-1+k+...) =


k-1
å
n = 0 
enxn+
å
n ³ k 
æ
è
en+(n-k+1)en-k+1ck+


Rn-k+1(e1,...,en-k) ö
ø
xn,
where Rn-k+1 = Rn-k+1(e1,...,en-k) is a polynomial in e1,...,en-k, and R1 = 0. Hence, a satisfies the boundary condition if and only if



å
n ³ 0 
æ
è
n
å
r = 0 
an-rer ö
ø
xn = em k-1
å
n = 0 
enxn+


em
å
n ³ k 
(en+(n-k+1)en-k+1ck+Rn-k+1) xn.
Comparing coefficients on both sides we derive for n = 0 that a0 = em, since e0 = 1, hence m = lna0. Then for 1 £ n < k the coefficient an = 0, since


n-1
å
r = 0 
an-rer+a0en = emen,
which is equivalent to


an+ n-1
å
r = 1 
an-rer = 0,
hence recursively we get


an = - n-1
å
r = 1 
an-rer = 0.
Finally, for n ³ k, write n as k+j for j ³ 0. Then the condition


a0ek+j+ k+j
å
s = k 
asek+j-s =


em(ek+j+(j+1)ej+1ck+Rj+1)
reduces to


j
å
r = 0 
ak+j-rer = em((j+1)ej+1ck+Rj+1),
from which ej+1 can uniquely be determined by


ej+1 =
e-m j
å
r = 0 
ak+j-rer-Rj+1

(j+1)ck
.
      [¯]

In order to deal with the general form of a let Pn,k(s,x) be given by


Pn,k(s,x): = æ
è
exp ó
õ
s

0 
p(s,x)nds ö
ø
k
 
=


exp æ
è
k ó
õ
s

0 
p(s,x)nds ö
ø
for 1 £ n < k. (The last equality is a consequence of the identity


(expF(y))k = exp(kF(y)),
holding for the formal series exp and the formal binomial series, with ord F(y) ³ 1.) Then P(s,x) = Õn = 1k-1Pn,kn(s,x).

Lemma 8 Let (p(s,x))s Î C be an analytic iteration group of the second type, k Î C and assume that 1 £ n < k. Then Pn,k(s,x) = 1+ksxn+....

Proof. Computing the first coefficients we get


p(s,x)n = xn+ æ
ç
è
n
1
ö
÷
ø
cksxn-1+k+...
hence


k ó
õ
s

0 
p(s,x)nds = ksxn+...
and consequently Pn,k(s,x) is of the given form.       [¯]

Standard computations can be used in order to prove

Lemma 9 Writing the series P(s,x), which is the product Õn = 1k-1Pn,kn(s,x), in the form


P(s,x) =
å
n ³ 0 
pn(s)xn,
then


pn(s) = ì
ï
í
ï
î
1,
n = 0
k1s,
n = 1
kns+qn(k1,...,kn-1,s),
2 £ n < k
where qn(k1,...,kn-1,s) is a polynomial in k1, ..., kn-1 and s. From this explicit form of pn(s) for 1 £ n < k it is possible to determine the vector of parameters (k1,...,kk-1) of a given polynomial P(s,x) in a unique way.

Already at the very beginning of this article we realized that a0(s) = ems, hence a0(1) = em = a0. Consequently it is enough and also easier to adjust [^(a)](s,x): = e-msa(s,x) to the boundary condition [^(a)](1,x) = [^a](x): = e-ma(x). The main idea is formulated in the next

Lemma 10 Let [^a](x) = 1+ån ³ n0[^a]nxn for 1 £ n0 < k. Then there exists exactly one Pn,k(s,x), such that


^
a
 
(x)

Pn,k(1,x)
º 1 mod ord ( n0+1).

Proof. When we choose n = n0 and k = [^a]n0, then it is clear from Lemma 3.7 that [^a](x) º Pn,k(1,x) mod ord ( n0+1). In order to prove that Pn,k(s,x) is uniquely defined, assume that there exists a series Pn¢,k¢(s,x), such that [^a](x) º Pn¢,k¢(1,x) mod ord (n0+1), then ord ([^a](x)- Pn¢,k¢(1,x)) ³ n0+1. Hence Pn¢,k¢(1,x) starts with 1+[^a]n0xn0. Consequently n¢ = n0 = n and k¢ = [^a]n0 = k by Lemma 3.7. Hence Pn¢,k¢(s,x) = Pn,k(s,x).       [¯]

From this lemma it is obvious that


^
a
 
(x) º Pn0,[^a]n0(1,x) mod ord ( n0+1).

Now we can adjust the general solution a given in the last part of Theorem 2.6 to the boundary condition.

Theorem 10 Let [^a](x) = 1+ån ³ 1[^a]nxn be a given formal power series of order 0 and assume that p is an analytic iteration group of the second type. Then there exists exactly one solution


^
a
 
(s,x) = P(s,x) E(p(s,x))
E(x)
of (Co1) with E(x) = 1+e1x+..., which also satisfies the boundary condition [^(a)](1,x) = [^a](x).

Proof. According to Lemma 3.9 there exists exactly one P1,k1(s,x), such that


^
a
 
(x)

P1,k1(1,x)
º 1 mod ord 2.
Assume that recursively for 1 £ n < k-1 we found uniquely defined Pn,kn(s,x), such that


^
a
 
(x)

P1,k1(1,x)¼Pn,kn(1,x)
º 1 mod ord ( n+1),
then by Lemma 3.9 there exists exactly one series Pn+1,kn+1(s,x), such that


^
a
 
(x)

P1,k1(1,x)¼Pn+1,kn+1(1,x)
º 1 mod ord ( n+2)
holds. Hence we end up with


^
a
 
(x)

P1,k1(1,x)¼Pk-1,kk-1(1,x)
º 1 mod ord k,
where P1,k1(s,x), ..., Pk-1,kk-1(s,x) are uniquely determined. From Theorem 3.6 we deduce the unique existence of a formal power series E(x) = 1+e1x+..., such that


^
a
 
(x)

P1,k1(1,x)¼Pk-1,kk-1(1,x)
= E(p(x))
E(x)
.
Thus [^a](x) can be written as


^
a
 
(x) = k-1
Õ
n = 1 
Pn,kn(1,x) E(p(x))
E(x)
,
where E(x) and Pn,kn(s,x) are uniquely determined for 1 £ n < k. From Lemma 3.8 it follows that there is exactly one vector of parameters of P(s,x): = Õn = 1k-1Pn,kn(s,x), namely (k1,...,kk-1), hence


^
a
 
(s,x): = P(s,x) E(p(s,x))
E(x)
is also uniquely determined by [^a](x). It is a solution of (Co1) and it satisfies the boundary condition, which finishes the proof.       [¯]

Summarizing, we found the following result: To any given formal power series a(x) of order 0 and any analytic iteration group p of the second type there exist solutions


a(s,x) = emsP(s,x) E(p(s,x))
E(x)
(15)
of (Co1) with E(x) = 1+e1x+..., which also satisfy the boundary condition a(1,x) = a(x).

Theorem 11 Assume that p is an analytic iteration group of second type. Let a(x) and b(x) be given formal power series, ord a(x) = 0, and let a be a solution of (Co1) of the form (15), which satisfies the boundary condition (B2).

  1. If a0 ¹ 1, then there exists exactly one


    b(s,x) =


    emsP(s,x)E(p(s,x))[F(x)-e-ms F(p(s,x))
    P(s,x)
    ],
    which satisfies together with a the cocycle equation (Co2) and the boundary condition b(1,x) = b(x).

  2. Assume that a0 = 1. If a(x) = x, let m0 = k, otherwise let m0 be the smallest element in { n Î N | an ¹ 0} and let n0: = min{ m0,k} .

    If n0 ¹ k-1, or kk-1-n ck ¹ 0 for all n Î N, then there exists exactly one b of the above form, which satisfies together with a the cocycle equation (Co2) and the boundary condition b(1,x) = b(x), if and only if bn = 0 for all 0 £ n < n0.

    If n0 = k-1 and kk-1 = n1ck for n1 Î N, there are two different cases to be considered.

    If m ¹ 0, then there exists a series b of the above form, which satisfies together with a the cocycle equation (Co2) and the boundary condition b(1,x) = b(x), if and only if bn = 0 for all 0 £ n < n0 and bn1+k-1 satisfies a condition which is implicitly given in the proof.

    If m = 0, then there exists a series


    b(s,x) = E(p(s,x))P(s,x)[F(x)-


    F(p(s,x))
    P(s,x)
    +ln1+n0¢¢ ó
    õ
    s

    0 
    p(s,x)n1+k-1
    P(s,x)
    ds ù
    ú
    û
    ,
    which satisfies together with a the cocycle equation (Co2) and the boundary condition b(1,x) = b(x), if and only if bn = 0 for all 0 £ n < k-1.

    In these last two situation, however, b is not uniquely determined.

Proof. Writing b as indicated in the first part of this theorem, the condition b(x) = b(1,x) is equivalent to


b(x)
E(p(x))
= emP(1,x)F(x)-F(p(x)).
(16)
From the proof of Theorem 3.6 we know that


E(p(x)) =
å
n ³ 0 
enxn+
å
n ³ k 
æ
è
(n-k+1)en-k+1ck+


Rn-k+1 ö
ø
xn º
å
n ³ 0 
enxn mod ord k.
If we denote b(x)/E(p(x)) by ån ³ 0[b\tilde]nxn , then


æ
è

å
n ³ 0 
~
b
 

n 
xn ö
ø
E(p(x)) =
å
n ³ 0 
bnxn.
Hence



å
n ³ 0 
bnxn º
å
n ³ 0 
æ
è
n
å
r = 0 
~
b
 

r 
en-r ö
ø
xn mod ord k,
and the coefficients bn are uniquely determined by the [b\tilde]n for 0 £ n < k.

Using the notation of Lemma 3.8 for the coefficients of P(s,x), condition (16) can be written as



å
n ³ 0 
~
b
 

n 
xn =


em æ
è

å
n ³ 0 
pn(1)xn ö
ø
æ
è

å
n ³ 0 
fnxn ö
ø
-
å
n ³ 0 
fnp(x)n =


em
å
n ³ 0 
æ
è
n
å
r = 0 
pr(1)fn-r ö
ø
xn-
å
n ³ 0 
fnxn-



å
n ³ k 
((n-k+1)fn-k+1ck+Sn-k+1) xn,
where Sn-k+1(f0,...,fn-k) is a polynomial in f0, ..., fn-k. Comparing coefficients yields


~
b
 

n 
= em æ
è
fn + n
å
r = 1 
pr(1)fn-r ö
ø
- fn
(17)
for 0 £ n < k, and


~
b
 

n 
= em æ
è
fn + n
å
r = 1 
pr(1)fn-r ö
ø
- fn -


(18)


(n-k+1)fn-k+1ck-Sn-k+1
for n ³ k.

First we assume that a0 = em ¹ 1. Then fn is uniquely determined by


fn =
~
b
 

n 
-em n
å
r = 1 
pr(1)fn-r

em-1
for 0 £ n < k, and


fn = (em-1)-1 æ
è
~
b
 

n 
-em n
å
r = 1 
pr(1)fn-r+


(n-k+1)fn-k+1ck+ Sn-k+1)
for n ³ k.

Next, assume that a0 = em = 1. From the definition of n0 we deduce that kn = 0 for 1 £ n < n0, and by Lemma 3.8 also pn(s) = 0 for 1 £ n < n0. And if moreover n0 < k, then pn0(s) = kn0s and kn0 ¹ 0. For n = 0 we deduce that [b\tilde]0 = 0 and recursively [b\tilde]n = 0 for all 1 £ n < n0. Hence, bn = 0 for 0 £ n < n0.

If n0 < k, then (17) means


~
b
 

n 
= n
å
r = n0 
pr(1)fn-r = kn0fn-n0 + n
å
r = n0+1 
pr(1)fn-r
for n0 £ n < k. Hence fn-n0 is uniquely determined by


fn-n0 =
~
b
 

n 
- n
å
r = n0+1 
pr(1)fn-r

kn0
for n0 £ n < k.

Finally assume that n ³ k. If n0 < k-1 (which is equivalent to n-n0 > n-k+1), then from (18) we deduce that [b\tilde]n equals


n
å
r = n0 
pr(1)fn-r-(n-k+1)fn-k+1ck-Sn-k+1,
hence


kn0fn-n0 = ~
b
 

n 
- n
å
r = n0+1 
pr(1)fn-r+


(n-k+1)fn-k+1ck +Sn-k+1,
which allows to determine fn-n0.

If n0 = k-1, then


~
b
 

n 
= kk-1fn-k+1 + n
å
r = k 
pr(1)fn-r-


(n-k+1)fn-k+1ck-Sn-k+1 =


(kk-1-(n-k+1)ck)fn-k+1+ n
å
r = k 
pr(1)fn-r- Sn-k+1.
If kk-1 ¹ mck for all m Î N, then fn-k+1 is uniquely determined by


fn-k+1 =
~
b
 

n 
- n
å
r = k 
pr(1)fn-r+Sn-k+1

kk-1-(n-k+1)ck
(19)
for all n ³ k.

The last case to be considered is the computation of fn for n ³ k, where a0 = 1, n0 = k-1 and kk-1 = n1ck for n1 Î N. First we assume that m ¹ 0. Then [b\tilde]n for n = n1+k-1, which is equivalent to n-k+1 = n1, must satisfy


~
b
 

n1+k-1 
= n1+k-1
å
r = k 
pr(1)fn1+k-1-r- Sn1 ,
(20)
such that b(s,x) can be adjusted to (B2). Hence bn1+k-1 must satisfy a corresponding condition. The coefficient fn1 of F(x) can be arbitrarily chosen in C, the coefficients fn-k+1 for n ³ k and n ¹ n1+k-1 are uniquely determined by (19).

If m = 0, then according to the last part of Theorem 2.8 we are allowed to add


E(p(s,x))P(s,x) ln1+n0¢¢ ó
õ
s

0 
p(s,x)n1+n0
P(s,x)
ds
to b(s,x), where ln1+n0¢¢ can be chosen in C. By doing this it is possible to skip the additional condition (20) for [b\tilde]n1+k-1. In order to compute the integral from above we derive


p(s,x)n1+n0 =


xn1+k-1+(n1+k-1)cksxn1+2(k-1)+...
and


p(s,x)n1+n0
P(s,x)
=


xn1+k-1+(n1+k-1)cksxn1+2(k-1)+...
1+kk-1sxk-1+...
=


xn1+k-1+((n1+k-1)ck-kk-1)sxn1+2(k-1)+... ,
hence


ó
õ
s

0 
p(s,x)n1+n0
P(s,x)
ds =


sxn1+k-1+ n0ck
2
s2 xn1+2(k-1)+... .
Consequently


P(1,x) ln1+n0¢¢ ó
õ
1

0 
p(s,x)n1+n0
P(s,x)
ds º


ln1+k-1xn1+k-1 mod ord (n1+k) .
We indicate this formal power series by



å
n ³ n1+k-1 
qnxn.

For k £ n < n1+k-1 the coefficient fn-k+1 is uniquely determined by (19). For n = n1+k-1 we determine ln1+k-1¢¢ by


ln1+k-1¢¢ = ~
b
 

n1+k-1 
- n1+k-1
å
r = k 
pr(1)fn1+k-1-r+Sn1
and fn1 can be chosen arbitrarily in C. Finally, for n > n1+k-1 the coefficients fn-k+1 is uniquely given by


fn-k+1 =
~
b
 

n 
- n
å
r = k 
pr(1)fn-r+Sn-k+1-qn

kk-1 -(n-k+1)ck
.
      [¯]

In the case a0 = 1 the condition bn = 0 for all 0 £ n < n0 is also a necessary condition for the existence of a solution j(x) of (L). This fact is shown in the next

Lemma 11 Let p(x) = x+ckxk+... for k ³ 2 and ck ¹ 0, and assume that a(x) = 1 (then set m0: = k) or a(x) = 1+am0xm0+... for 1 £ m0 and am0 ¹ 0. If j(x) Î C [[x]] is a solution of (L), then bn = 0 for 0 £ n < n0: = min{ m0,k} .

Proof. Elementary computations yield


j(p(x)) =
å
n ³ 0 
jnp(x)n =



å
n ³ 0 
jn(xn+nckxn-1+k+...) º



å
n ³ 0 
jnxn mod ord k,
and


a(x)j(x)+b(x) =
å
n ³ 0 
æ
è
n
å
r = 0 
arjn-r+bn ö
ø
xn =



å
n ³ 0 
æ
è
jn+ n
å
r = m0 
arjn-r+bn ö
ø
xn.
Hence, comparing coefficients of xn on the left and on the right side of (L) yields


jn = jn+bn
for 0 £ n < n0, thus bn = 0.       [¯]

4  Solution of the problem of covariant embeddings in certain special cases

In this section we give in Theorem 4.1 a necessary condition that a given linear functional equation (L) (with a non-empty set of solutions) has an embedding with respect to a given analytic iteration group of p(x). In Corollary 4.2 we present, as a consequence, a rather large class of such embeddings. However, there remain some special cases of solutions (a,b) of the system ((Co1),(Co2)) and the boundary conditions corresponding to (L) where the existence of an embedding is still open.

Theorem 12 Assume that the linear functional equation (L) has a solution j(x) Î C [[x]], and let (p(s,x))s Î C be an analytic iteration group of p(x). Furthermore, assume that a satisfies (Co1) and the two boundary conditions (B1) and (B2). If there exists exactly one b, which also satisfies (B1) and (B2), such that (a,b) is a solution of (Co2), then there exists an embedding of (L) with respect to the iteration group (p(s,x))s Î C.

Proof. Let j be a solution of (L). Then Fj(s,x) defined by


Fj(s,x): = j(p(s,x))-a(s,x)j(x)
satisfies both Fj(0,x) = j(x)-1j(x) = 0 and Fj(1,x) = j(p(x))-a(x)j(x) = b(x). Furthermore, the pair (a,Fj) satisfies (Co2), since


Fj(t+s,x) = j(p(t+s,x))-a(t+s,x)j(x) =


j(p(t,p(s,x)))-a(s,x)a(t,p(s,x))j(x) =


Fj(t,p(s,x))+a(t,p(s,x))j(p(s,x))-


a(s,x)a(t,p(s,x))j(x) =


Fj(t,p(s,x))+a(t,p(s,x))·


·[Fj(s,x)+a(s,x)j(x)-a(s,x)j(x)] =


Fj(t,p(s,x))+a(t,p(s,x))Fj(s,x).
In other words, Fj satisfies the same conditions as b, i.e. (B1), (B2) and together with a the cocycle equation (Co2). Hence, since there exists exactly one b with these properties, Fj(s,x) = b(s,x) for all s Î C and all solutions j(x) of (L).       [¯]

Combining this result with Theorem 3.5 and Theorem 3.11 we get

Corollary 2 If p(s,x) = elsx is an analytic iteration group of the first type, and el is not a complex root of 1, then there exists an embedding of (L) with respect to the iteration group p.

If p(s,x) = x+cksxk+... with k ³ 2 and ck ¹ 0 is an analytic iteration group of the second type, and a0 ¹ 1 or n0 < k-1 or ak-1 ¹ nck for all n Î N, then there exists an embedding of (L) with respect to the iteration group p.

If p(x) = elx, where el ¹ 1 is a complex root of 1, and p(x) does not have an embedding in an analytic iteration group, then there exists no covariant embedding of the linear functional equation. If p(x) has an embedding, then it is still open whether there exists a covariant embedding of (L). In addition to this the embedding problem is also still open for analytic iteration groups p of second type, when p(s,x) = x+cksxk+... with k ³ 2 and ck ¹ 0 and a(x) = 1+ak-1xk-1+... with ak-1 ¹ 0 and ak-1 = n1ck for n1 Î N.

References

[Cartan, 1963]
H. Cartan. Elementary theory of analytic functions of one or several complex variables. Addison- Wesley Publishing Company, Reading (Mass.), Palo Alto, London, 1963.

[Cartan, 1966]
H. Cartan. Elementare Theorie der analytischen Funktionen einer oder mehrerer komplexen Veränderlichen, volume 112/112a. BI-Hochschultaschenbücher, Mannheim, Wien etc., 1966.

[Guzik, 1999]
G. Guzik. On embedding of a linear functional equation. Wy\.z Szkoa Ped. Kraków Rocznik Nauk.-Dydact. Prace Matematyczne, 16:23-33, 1999.

[Guzik, 2000]
G. Guzik. On continuity of measurable cocycles. Journal of Applied Analysis, 6(2):295-302, 2000.

[Guzik, 2001]
G. Guzik. On embeddability of a linear functional equation in the class of differential functions. Grazer Mathematische Berichte, 344:31-42, 2001.

[Henrici, 1977]
P. Henrici. Applied and computational complex analysis. Vol. I: Power series, integration, conformal mapping, location of zeros. John Wiley & Sons, New York etc., 1977.

[Kuczma, 1968]
M. Kuczma. Functional equations in a single variable, volume 46 of Monografie Math. Polish Scientific Publishers, Warsaw, 1968.

[Kuczma et al., 1990]
M. Kuczma, B. Choczewski, and R. Ger. Iterative Functional Equations, volume 32 of Encyclopedia of Mathematics and its Applications. Cambridge University Press, 1990.

[Moszner, 1999]
Z. Moszner. Sur le prolongement covariant d'une équation linéare par rapport au groupe d'itération. Sitzungsberichte ÖAW, Math.-nat. Kl. Abt. II, 207:173-182, 1999.

[Reich, 1998]
L. Reich. 24. Remark in The Thirty-fifth International Symposium on Functional Equations, September 7-14, 1997, Graz-Mariatrost, Austria. Aequationes Mathematicae, 55:311-312, 1998.

[Reich & Schwaiger, 1977]
L. Reich and J. Schwaiger. Über einen Satz von Shl. Sternberg in der Theorie der analytischen Iterationen. Monatshefte für Mathematik, 83:207-221, 1977.

[Scheinberg, 1970]
St. Scheinberg. Power Series in One Variable. Journal of Mathematical Analysis and Applications, 31:321-333, 1970.


Footnotes:

1Supported by the Fonds zur Förderung der wissenschaftlichen Forschung P14342-MAT.


File translated from TEX by TTH, version 2.79.
On 18 Apr 2001, 16:28.