1 Introduction: Invariance Under Binomial Thinning in Power Series Families

Functional equations we analyze in this paper arise naturally in an invariance problem involving Poisson-type random probability measures, known under the nickname throwing stones and collecting bones. The problem has two basic ingredients: the binomial thining operator and the power series family of probability distributions.

  1. 1.

    Binomial thinning: Let \(\mathcal P({\mathbb N})\) be the set of probability measures with supports in \({\mathbb N}=\{0,1,\ldots \}\). For every \(p\in [0,1]\) the binomial thining operator \(T_p\) is defined as follows:

    $$\begin{aligned} \mathcal P({\mathbb N})\ni \mu \mapsto T_p(\mu )\in \mathcal P({\mathbb N}) \end{aligned}$$

    and \(T_p(\mu )\) is the probability distribution of

    $$\begin{aligned} \tilde{K}:=\sum _{n=0}^K\,I_n, \end{aligned}$$
    (1)

    where the sequence \((I_n)_{n\ge 1}\) of independent random variables with the same Bernoulli distribution \(\mathrm {Ber}(p):=(1-p)\delta _0+p\delta _1\) (additionally we denote \(I_0=0\)) and the random variable K with distribution \(\mu \) (defined on some probability space \((\Omega ,\mathcal F,\mathbb P)\)) are independent. Here and in the sequel by \(\delta _x\) we denote the Dirac measure at x. In particular, \(T_0(\mu )=\delta _0\) and \(T_1(\mu )=\mu \). This operator was introduced in [13] to establish discrete versions of stability and selfdecomposability of probability measures. Since then binomial thinning operator and its extensions have been intensively studied in various probabilistic contexts (a prominent example being the time series theory). In particular, very recently [3] (referred to by BR in the sequel) used the thinning operator to model Poisson-type random point processes restricted to a subset of the original state space.

  2. 2.

    Power series family: Let \(\underline{a}=(a_k)_{k\ge 0}\) be a sequence of nonnegative numbers with \(a_0=1\) such that the set

    $$\begin{aligned} \Theta _{\underline{a}}=\left\{ \theta \ge 0:\,\varphi (\theta ):=\sum _{k\ge 0}\,a_k\theta ^k<\infty \right\} \end{aligned}$$

    has a non-empty interior (actually \(\Theta _{\underline{a}}\) is a convex set). Then

    $$\begin{aligned} \mu _{\underline{a},\theta }=\sum _{k\ge 0}\,\tfrac{a_k\theta ^k}{\varphi (\theta )}\,\delta _k \end{aligned}$$

    is a probability measure called a power series distribution generated by \(\underline{a}\) with the parameter \(\theta \in \Theta _{\underline{a}}\). The power series family generated by \(\underline{a}\) is defined as

    $$\begin{aligned} \mathcal {PSF}(\underline{a})=\{\mu _{\underline{a},\theta }:\,\theta \in \Theta _{\underline{a}}\}. \end{aligned}$$

The problem lies in identification of power series families which are invariant under binomial thinning, i.e. one searches for \(\mathcal {PSF}(\underline{a})\) satisfying

$$\begin{aligned} T_p(\mathcal {PSF}(\underline{a}))\subset \mathcal {PSF}(\underline{a}) \end{aligned}$$

for some \(p\in (0,1)\).

Equivalently, we want to describe all sequences \(\underline{a}\) of nonnegative numbers with \(a_0=1\) and with \(\Theta _{\underline{a}}\) of non-empty interior, such that there exists \(p\in (0,1)\) and a function \(h_p:\Theta _{\underline{a}}\rightarrow \Theta _{\underline{a}}\) which satisfy the condition

$$\begin{aligned} \forall \,\theta \in \Theta _{\underline{a}}\quad \left( \, K\sim \mu _{\underline{a},\theta }\quad \Rightarrow \quad \tilde{K}\sim \mu _{\underline{a},h_p(\theta )}\,\right) , \end{aligned}$$
(2)

where \(\tilde{K}\) is defined in (1). (If \(\mu \) is the probability distribution of a random variable X we write \(X\sim \mu \).)

Probability generating function is a convenient tool to analyze this problem. Recall that the probability generating function \(\psi _{\mu }\) of \(\mu =\sum _{k\ge 0}\,p_k\delta _k\in \mathcal {P}({\mathbb N})\) is defined by \(\psi _{\mu }(s)=\sum _{k\ge 0}\,s^kp_k\) on a domain \(U\supset [-1,1]\). In particular,

$$\begin{aligned} \psi _{\mathrm {Ber}(p)}(s)=ps+q,\quad s\in {\mathbb R},\quad \text{ where } \;q=1-p, \end{aligned}$$

and

$$\begin{aligned} \psi _{\mu _{\underline{a},\theta }}(s)=\sum _{k\ge 0}\,s^k\,\tfrac{a_k\theta ^k}{\varphi (\theta )}=\tfrac{\varphi (s\theta )}{\varphi (\theta )},\qquad |s|\theta \in \Theta _{\underline{a}},\;\theta \in \Theta _{\underline{a}}, \end{aligned}$$

where \(\varphi \) is defined on \(\Theta _{\underline{a}}\cup (-\Theta _{\underline{a}})\) by analytical extension.

Let \(\nu \) be the distribution of \(\tilde{K}\) defined in (1) with \(K\sim \mu _{\underline{a},\theta }\) for \(\theta \in \Theta _{\underline{a}}\). Then using conditioning with respect to K and independence of \(K,I_1,I_2,\ldots \) we get

$$\begin{aligned} \psi _{\nu }(s)=\mathbb E\,s^{\tilde{K}}=\mathbb E\,s^{\sum _{n=0}^K\,I_n}=\mathbb E\,(\psi _{\mathrm {Ber}(p)}(s))^K=\psi _{\mu _{\underline{a},\theta }} \left( \psi _{\mathrm {Ber}(p)}(s)\right) =\tfrac{\varphi ((ps+q)\theta )}{\varphi (\theta )}, \end{aligned}$$

if only \(\theta ,\,|ps+q|\theta \in \Theta _{\underline{a}}\).

By (2) we have \(\nu =\mu _{\underline{a},h_p(\theta )}\), i.e. for \(p\in (0,1)\) we get the equation

$$\begin{aligned} \tfrac{\varphi ((ps+q)\theta )}{\varphi (\theta )}=\tfrac{\varphi (sh_p(\theta ))}{\varphi (h_p(\theta ))} \end{aligned}$$
(3)

for s and \(\theta \) satisfying \(\theta ,\,|ps+q|\theta ,\,|s|h_p(\theta )\in \Theta _{\underline{a}}\). Since \(q\theta \in \Theta _{\underline{a}}\), upon inserting \(s=0\) in (3), we get (note that \(\varphi (0)=1\))

$$\begin{aligned} \varphi (h_p(\theta ))=\tfrac{\varphi (\theta )}{\varphi (q\theta )},\quad \theta \in \Theta _{\underline{a}}. \end{aligned}$$

Consequently, (3) can be rewritten as

$$\begin{aligned} \varphi ((ps+q)\theta )=\varphi (q\theta )\,\varphi (sh_p(\theta )). \end{aligned}$$
(4)

Then, upon changing variables \(u:=ps\theta \), \(v:=q\theta \) the equation (4) yields

$$\begin{aligned} \varphi (u+v)=\varphi (v)\,\varphi (u\rho (v)) \end{aligned}$$
(5)

on the proper domain for variables u and v (actually, this domain contains a neighbourhood of zero for u and a right neighbourhood of zero for v), where \(\rho (0)=1\) and \(\rho (v)=\tfrac{q}{pv}h_p\left( \tfrac{v}{q}\right) \), \(v>0\). Since \(\varphi (0)=1\) one can apply the logarithm to both sides of (5) for u in a two-sided neighbourhood of zero and v in the right-neighbourhood of zero which leads to an additive version of this equation (5). Such equation, referred to by the modified Cauchy equation in BR, has been recently solved in that paper. We quote now this result in extenso:

Lemma 1.1

(BR, Lemma 1) Assume that f(t) is twice differentiable in some neighbourhood of the origin, satisfies \(f(0)=0\) and \(f'(0)>0\) as well as

$$\begin{aligned} f(s+t)-f(s)=f(h(s)t), \end{aligned}$$

where h(s) is t free. Then f is of the form \(f(t)=At\) or \(f(t)=B\log (1+At)\) for some \(A,B\ne 0\). Moreover \(h(s)=f'(s)/f'(0)\).

In BR Lemma 1.1 is used to identify Poisson, binomial and negative binomial probability distributions as the only power series families which are invariant under binomial thinning (see Theorem 2 and its proof in BR).

We are interested in the above equation as well as its ”dual” \(f(s+g(s)t)=f(s)+f(t)\). Instead of a neighbourhood of zero we consider domains: \([0,\infty )\) in Section 3 and V, a vector space, in Section 4. We assume minor or no regularity conditions on f and consider several cases of image spaces of f: a unital magma, the real line and a linear topological space. No regularity conditions whatever are imposed on the unknown functions h and g. Our results complete to some extent [6] where all solutions of the ”dual” equation are determined in the case when f maps a linear space (real or complex) into a semigroup and g satisfies some regularity conditions. In Section 2 we give preliminaries on the functional equations we are interested in.

2 Cauchy–Goła̧b–Schinzel Equations

Let \((M,+)\) be a magma, i.e. M is a set equipped with a binary operation \(+:M\times M\rightarrow M\). Consider \(U=[0,\infty )\) or \(U=V\), a vector space over a field \(\mathbb F\). For unknown functions \(f:U\rightarrow M\) and \(g,h:U\rightarrow W\), where \(W=U\) in case \(U=[0,\infty )\) and \(W=\mathbb F\) in case \(U=V\), we consider equations

$$\begin{aligned} f(s+t)=f(s)+f(h(s)t),\qquad s,t\in U, \end{aligned}$$
(6)

and

$$\begin{aligned} f(s+g(s)t)=f(s)+f(t),\qquad s,t\in U. \end{aligned}$$
(7)

We would like to identify all solutions (fh) of (6) and (fg) of (7).

If \(f=g\) (thus \(M=W\)), then (7) becomes an additive form of the Goła̧b–Schinzel equation, see [1], pp. 132-135, [2], pp. 311-319 and the survey paper [5]. For more recent contributions on the Goła̧b–Schinzel equation and its generalizations consult e.g. [4, 7,8,9] and [11]. In particular, the latter paper reveals yet another probabilistic (stable laws and random walks) connection of the Goła̧b–Schinzel equation, treated there as a disguised form of the Goldie equation. If \(g\equiv 1\) or \(h\equiv 1\), then (6) and (7) become the same standard Cauchy equation. Hence we call (7) as well as (6) the Cauchy–Goła̧b–Schinzel (CGS) equations.

We consider unital magma M (i.e. M has a neutral element, denoted by \(\mathbf{0}\) throughout the paper) with the two-sided cancelation property. To avoid trivialities we assume that \(f\not \equiv \mathbf{0}\).

Note that for f which solves either (6) or (7) we have

$$\begin{aligned} f(0)=\mathbf{0}. \end{aligned}$$
(8)

Remark 2.1

If (fg) solves (7), then \(\mathrm {Ker}(g):=\{s\in U:\,g(s)=0\}=\emptyset \). Assume not, i.e. \(g(s_0)=0\) for some \(s_0\in U\). Then (7) implies \(f(s_0)=f(s_0+g(s_0)t)=f(s_0)+f(t)\) for any \(t\in U\). Hence \(f\equiv \mathbf{0}\), a contradiction.

Remark 2.2

If (fh) solves (6) and \(\mathrm {Ker}(h)=\emptyset \), then (fg) with \(g=1/h\) solves (7). In the opposite direction, if (fg) solves (7), then (fh) with \(h=1/g\) (being well-defined by Remark 2.1) solves (6).

Remark 2.3

Let (fh) solves (6) for \(U=V\), a vector space. Then \(\mathrm {Ker}(h)=\emptyset \).

Assume not, i.e. \(h(s_0)=0\) for some \(s_0\in V\). Then (6) together with (8) imply \(f(s_0+t)=f(s_0)\) for every \(t\in V\). That is, f is a constant function. By (8) we get a contradiction with \(f\not \equiv \mathbf{0}\).

When \(U=[0,\infty )\), while considering (6) it is convenient to distinguish two cases with respect to the form of the kernel of h:

  1. I

    \(\mathrm {Ker}(h)\ne \emptyset \)

  2. II

    \(\mathrm {Ker}(h)=\emptyset \)

3 CGS Equations on \(U=[0,\infty )\)

3.1 \(\mathbf {\mathrm {\mathbf {Ker}}(h)\ne \emptyset }\)

Throughout this section we assume that \(U=[0,\infty )\) and that \((M,+)\) is a unital magma with the two-sided cancelation property.

Theorem 3.1

Assume that \(f:[0,\infty )\rightarrow M\) is non-zero, \(h:[0,\infty )\rightarrow [0,\infty )\), \(\mathrm {Ker}(h)\ne \emptyset \) and

$$\begin{aligned} s_0=\inf \,\mathrm {Ker}(h). \end{aligned}$$

Then (fh) solves (6) on \([0,\infty )\) if and only if

  1. 1.

    either \(s_0=0\) and

    $$\begin{aligned} f(s)=\left\{ \,\begin{array}{ll} \mathbf{0},&{} \mathrm {for}\;s=0, \\ \mathbf{a}, &{} \mathrm {for}\;s\in (0,\infty ), \end{array}\right. \qquad h(s)=\left\{ \,\begin{array}{ll} b,&{} \mathrm {for}\;s=0, \\ 0, &{} \mathrm {for}\;s\in (0,\infty ), \end{array}\right. \end{aligned}$$
    (9)

    where \(\mathbf{a}\in M\setminus \{\mathbf{0}\}\) and \(b\in (0,\infty )\),

  2. 2.

    or \(s_0>0\) and

    $$\begin{aligned} f(s)=\left\{ \,\begin{array}{ll} \mathbf{0},&{} \mathrm {for}\;s\in [0,s_0), \\ \mathbf{a}, &{} \mathrm {for}\;s\in [s_0,\infty ), \end{array}\right. \qquad h(s)=\left\{ \,\begin{array}{ll} \tfrac{s_0}{s_0-s},&{} \mathrm {for}\;s\in [0,s_0), \\ 0, &{} \mathrm {for}\;s\in [s_0,\infty ), \end{array}\right. \end{aligned}$$
    (10)

    where \(\mathbf{a}\in M\setminus \{\mathbf{0}\}\).

Proof

It is easy to check that (fh) given in (9) and (10) solve (6).

Let (fh) solve (6). Then it follows from (6) and (8) that for any \(t\in [0,\infty )\),

$$\begin{aligned} f(s+t)=f(s)\qquad \text{ if } \text{ only } \;h(s)=0. \end{aligned}$$

Since such \(s\ge s_0\) can be chosen arbitrarily close to \(s_0\) we conclude that

$$\begin{aligned} f(s)=\mathbf{a}\in M,\qquad s\in (s_0,\infty ). \end{aligned}$$
(11)

If \(s\in [0,\infty )\) is such that \(h(s)\ne 0\), then we consider (6) for \(t>0\) such that \(h(s)t>s_0\) and \(s+t>s_0\). Then (11) yields \(\mathbf{a}=f(s)+\mathbf{a}\), whence \(f(s)=0\). Therefore,

$$\begin{aligned} h(s)\ne 0\quad \Rightarrow \quad f(s)=\mathbf{0},\qquad s\in [0,\infty ). \end{aligned}$$
(12)

Combining (11) and (12) we get

$$\begin{aligned} h(s)=0,\quad s\in (s_0,\infty )\qquad \text{ if } \text{ only } \; \mathbf{a}\ne \mathbf{0}. \end{aligned}$$
(13)

Consider now two cases.

  1. 1.

    \(s_0=0\): Then (11) together with (8) imply that \(\mathbf{a}\ne \mathbf{0}\). Consequently, f is as given in (9). By (13) we have \(h(s)=0\) for \(s>0\). From (6) for \(s=0\) and \(t>0\) we get \(\mathbf{a}=f(t)=f(h(0)t)\). Thus (8) implies \(h(0)>0\). Consequently, h is as given in (9).

  2. 2.

    \(s_0>0\): Then (12) gives \(f(s)=\mathbf{0}\) for \(s\in [0,s_0)\). From (6) for \(s=s_0\) and \(t>0\) such that \(h(s_0)t<s_0\) by (11) we get \(\mathbf{a}=f(s_0)\) which implies that \(\mathbf{a}\ne \mathbf{0}\). Consequently, f is as given in (10). By (13) and (12) we have \(h(s)=0\) for \(s\in [s_0,\infty )\). Let \(s\in [0,s_0)\). Then \(f(s+t)=f(h(s)t)\) for \(t\in [0,\infty )\). Referring to f as given in (10) we see that

    $$\begin{aligned} s+t<s_0\quad \Leftrightarrow \quad h(s)t<s_0. \end{aligned}$$

    Thus

    $$\begin{aligned} \tfrac{s_0}{s_0-s}\le h(s)<\tfrac{s_0}{s_0-s-\epsilon }\qquad \text{ for }\;\;\epsilon \in (0,s_0-s]. \end{aligned}$$

    By taking \(\epsilon \downarrow 0\) we obtain \(h(s)=\tfrac{s_0}{s-s_0}\) for \(s\in [0,s_0)\). Consequently, h is as given in (10).

\(\square \)

3.2 \(\mathbf {\mathrm {\mathbf {Ker}}(h)=\emptyset }\)

Let \(U=[0,\infty )\). As explained in Remark 2.2, (fh) solves (6) if and only if (fg) with \(g=1/h\) solves (7). We start with the case where f is injective; cf. [14]. Then we move on to the case when f is right continuous at a point.

3.2.1 A Magma Version

Throughout this section we assume that \((M,+)\) is a commutative unital magma with the (two-sided) cancelation property.

Theorem 3.2

Assume that \(f:[0,\infty )\rightarrow M\) is injective and \(g:[0,\infty )\rightarrow [0,\infty )\).

Then (fg) solves (7) if and only if

  1. 1.

    either \(g(1)=1\) and

    $$\begin{aligned} f\;\text{ is } \text{ additive }\qquad \text{ and }\qquad g\equiv 1, \end{aligned}$$
    (14)
  2. 2.

    or \(g(1)\ne 1\) and

    $$\begin{aligned} f(s)=a(\log (\alpha s+1))\qquad \text{ and }\qquad g(s)=\alpha s+1,\quad s\in [0,\infty ), \end{aligned}$$
    (15)

    where \(\alpha \in (0,\infty )\) and \(a:[0,\infty )\rightarrow M\) is an injective additive function.

Proof

It is easy to check that (fg) given in (14) and (15) solve (7). By commutativity at the right hand side of (7) and injectivity of f we conclude that

$$\begin{aligned} s+g(s)t=t+g(t)s,\quad s,t\in [0,\infty ). \end{aligned}$$

For \(t=1\) we get

$$\begin{aligned} g(s)=\alpha s+1,\quad s\ge 0, \end{aligned}$$
(16)

where \(\alpha =g(1)-1\) and \(\alpha \ge 0\) since g is positive. Consider two possible cases: \(g(1)=1\) and \(g(1)\ne 1\).

  1. 1.

    \(g(1)=1\): By (16) we have \(\alpha =0\) and \(g\equiv 1\). Then by (7) it follows that f is additive.

  2. 2.

    \(g(1)\ne 1\): By (16) we have \(\alpha >0\) and g is as given in (15). It follows from (16) that \(k:=\log (g):[0,\infty )\rightarrow [0,\infty )\) is a bijection (note that \(g(s)\ge 1\) for all \(s\in [0,\infty )\)). Therefore the formula \(a\circ k=f\) defines a function \(a:[0,\infty )\rightarrow M\). Clearly, a is injective and the Eq. (7) in terms of a assumes the form

    $$\begin{aligned} a(k(s))+a(k(t))=a(k(s+g(s)t)),\quad s,t\ge 0. \end{aligned}$$

    But g, see (16), satisfies \(g(s+g(s)t)=g(s)g(t)\), i.e. \(k(s+g(s)t)=k(s)+k(t)\). Therefore

    $$\begin{aligned} a(k(s))+a(k(t))=a(k(s)+k(t)),\quad s,t\ge 0. \end{aligned}$$

    Since k is bijective on \([0,\infty )\) we conclude that a is an additive function.

\(\square \)

3.2.2 Real and Vector Space Versions

We first consider the case of real-valued f.

Theorem 3.3

Assume that \(f:[0,\infty )\rightarrow {\mathbb R}\) is non-zero, right continuous at some point and \(g:[0,\infty )\rightarrow [0,\infty )\). Then (fg) solves (7) if and only if

  1. 1.

    either

    $$\begin{aligned} f(s)=as,\quad s\in [0,\infty )\qquad \text{ and }\qquad g\equiv 1, \end{aligned}$$

    where \(0\ne a\in {\mathbb R}\),

  2. 2.

    or

    $$\begin{aligned} f(s)=a\log (\alpha s+1)\qquad \text{ and }\qquad g(s)=\alpha s+1,\quad s\in [0,\infty ), \end{aligned}$$

    where \(\alpha >0\) and \(0\ne a\in {\mathbb R}\).

For the proof of Theorem 3.3 we use several auxiliary results which are considered first.

In Lemmas 3.4, 3.5, 3.6, 3.8 and 3.9 as well as in Proposition 3.7 we assume that a non-zero function \(f:[0,\infty )\rightarrow {\mathbb R}\) satisfies (7) with some \(g:[0,\infty )\rightarrow [0,\infty )\).

Lemma 3.4

If f is right continuous at some point, then it is a right continuous function.

Proof

Let f be right continuous at \(s_1\in [0,\infty )\). From (7) we have

$$\begin{aligned} f(s_1+g(s_1)t)-f(s_1)=f(t),\quad t\ge 0. \end{aligned}$$

Taking \(t\rightarrow 0^+\) we see that right continuity of f at \(s_1\) implies that f is right continuous at 0.

Fix arbitrary \(s>0\). Then by (7) for \(t\ge 0\) we have

$$\begin{aligned} f(s+t)-f(s)=f(h(s)t), \end{aligned}$$

where \(h=1/g\). Taking \(t\rightarrow 0^+\) we conclude that f is right continuous at s.

\(\square \)

Lemma 3.5

If f is right continuous, then \(g(s)\ge 1\) for every \(s\ge 0\).

Proof

Assume \(g(s_1)<1\) for some \(s_1\ge 0\). Then for \(\phi :[0,\infty )\rightarrow [0,\infty )\) defined by \(\phi (t)=g(s_1)t+s_1\),

$$\begin{aligned} \tilde{s}<\phi (t)<t\qquad \forall \,t>\tfrac{s_1}{1-g(s_1)}=:\tilde{s}. \end{aligned}$$

Hence \(t_n:=\phi ^{\circ n}(t)=g^n(s_1)t+s_1\sum _{k=0}^{n-1}\,g^k(s_1)\rightarrow \,\tilde{s}^+\) as \(n\rightarrow \infty \) for \(t>\tilde{s}\). Thus right continuity of f implies

$$\begin{aligned} f(\tilde{s})-f(t)=\lim _{n\rightarrow \infty }\,\left( f(t_n)-f(t)\right) . \end{aligned}$$
(17)

On the other hand for any \(k\ge 1\) we have \(t_k=\phi (t_{k-1})\) and thus (7) yields

$$\begin{aligned} f(t_k)=f(g(s_1)t_{k-1}+s_1)=f(t_{k-1})+f(s_1) \end{aligned}$$

whence

$$\begin{aligned} f(t_n)-f(t)=\sum _{k=1}^n\left( f(t_k)-f(t_{k-1})\right) = nf(s_1)\qquad \forall \,t>\tilde{s},\;n\ge 1. \end{aligned}$$

Thus, by (17), we obtain

$$\begin{aligned} f(\tilde{s})-f(t)=\lim _{n\rightarrow \infty }\,nf(s_1)\qquad \forall \,t>\tilde{s}. \end{aligned}$$

Therefore \(f(s_1)=0\) and \(f|_{(\tilde{s},\infty )}\equiv a:=f(\tilde{s})\). Taking now \(s,t>\tilde{s}\) in (7) we get \(a=0\).

Let \(s'=\inf \{s\ge 0:\,f|_{(s,\infty )}\equiv 0\}\). Assume \(s'>0\). Then for \(s\in (0,s')\) and \(t>s'\) in (7) we have \(0=f(t+g(t)s)=f(s)\) - a contradiction. Thus \(s'=0\). But this is impossible since \(f\not \equiv 0\).

\(\square \)

Lemma 3.6

If f is right continuous, then it is a continuous function.

Proof

It suffices to prove that f is left continuous at any \(s>0\). Fix arbitrary \(s>0\). Then by (7) we have

$$\begin{aligned} f(s-t)-f(s)=-f(th(s-t))\qquad \forall \,t\in [0,s], \end{aligned}$$

where \(h=1/g\). By Lemma 3.5 we have \(0<th(s-t)\le t\) for \(t\in (0,s]\). Therefore, since f is right continuous at 0, for \(t\rightarrow 0^+\) the left hand side turns to zero and the result follows.

\(\square \)

Combining Lemmas 3.4 and 3.6 we get the following result.

Proposition 3.7

If f is right continuous at some point, then it is a continuous function.

If f is monotone, then it has a countable set of points of discontinuity, i.e. in view of Proposition 3.7 any monotone f satisfying (7) is continuous. It appears that this implication can be reversed with the help of right upper and lower Dini derivatives \(D^+\) and \(D_+\).

Lemma 3.8

If f is continuous, then it is a monotone function.

Proof

There exists a sequence \((a_n)_{n\ge 1}\) in \((0,\infty )\) such that \(\lim _{n\rightarrow \infty }\,a_n=0\) and, see (8),

$$\begin{aligned} \lim _{n\rightarrow \infty }\,\tfrac{f(a_n)}{a_n}= D^+f(0). \end{aligned}$$

Then, by (7), for every \(s\ge 0\) we get

$$\begin{aligned} D^+f(s)\ge \lim _{n\rightarrow \infty }\,\tfrac{f\left( s+a_ng(s)\right) -f(s)}{a_ng(s)}=\lim _{n\rightarrow \infty }\,\tfrac{f\left( a_n\right) }{a_ng(s)}=\tfrac{D^+f(0)}{g(s)}. \end{aligned}$$

Similarly, there exists a sequence \((b_n)_{n\ge 1}\) in \((0,\infty )\) such that \(\lim _{n\rightarrow \infty }\,b_n=0\) and

$$\begin{aligned} D_+f(s)\le \lim _{n\rightarrow \infty }\,\tfrac{f\left( s+b_ng(s)\right) -f(s)}{b_ng(s)}=\lim _{n\rightarrow \infty }\,\tfrac{f\left( b_n\right) }{b_ng(s)}=\tfrac{D_+f(0)}{g(s)}. \end{aligned}$$

Therefore

$$\begin{aligned} D_+f(s)\le \tfrac{D_+f(0)}{g(s)}\le \tfrac{D^+f(0)}{g(s)}\le D^+f(s),\qquad s\ge 0. \end{aligned}$$

Consequently either \(D^+f(s)\ge 0\) for every \(s\ge 0\) or \(D_+f(s)\le 0\) for every \(s\ge 0\). From [10], Theorem 7.4.13 and its Corollary, it follows that f is either a non-decreasing or a non-increasing function.

\(\square \)

Finally we connect monotonicity of f with its injectivity.

Lemma 3.9

If f is monotone, then it is an injective function.

Proof

Suppose f is not injective. Then \(f(t_1)=f(t_2)\) for some \(t_1,t_2\ge 0\) such that \(t_1<t_2\). By (7), with arguments \(t_1\) and \(s_1=\tfrac{t_2-t_1}{g(t_1)}>0\) it follows that

$$\begin{aligned} f(s_1)=f(t_1+g(t_1)s_1)-f(t_1)=f(t_2)-f(t_1)=0. \end{aligned}$$

Since f is monotone and \(f(0)=0\) it follows that \(f|_{[0,s_1]}\equiv 0\). Therefore

$$\begin{aligned} s'=\sup \{r\ge 0:\,f|_{[0,r]}\equiv 0\}>0. \end{aligned}$$

For \(s,t\in [0,s')\) equation (7) implies \(f(s+g(s)t)=0\), whence \(s+g(s)t\le s'\). It means that \(g(s)\le \tfrac{s'-s}{t}\) for every \(s,t\in [0,s')\). Thus taking \(t\in (s'-s,s')\) we get \(g(s)<1\), \(s\in (0,s')\).

Since f is monotone then it is right continuous at a point in \([0,\infty )\) and, in view of Lemma’s 3.4 and 3.5, \(g(s)\ge 1\) for every \(s\ge 0\), a contradiction.

\(\square \)

Proof

[Proof of Theorem 3.3] By Proposition 3.7 and Lemma 3.8 function f is monotone and it follows from Lemma 3.9 that f is injective. Thus by referring to Theorem 3.2 we conclude the proof since an additive and monotone function is linear - see e.g. [1], Ch. 2.1.1.

\(\square \)

Theorem 3.3 can be extended to f assuming values in a real topological vector space X with dual \(X^*\) which separates points on X, i.e. for every \({\mathbf{x}}\in X\setminus \{\mathbf{0}\}\) there exists an \(x^*\in X^*\) such that \(x^*\mathbf{x}\ne 0\). Consequently, for \(\mathbf{a},\,\mathbf{b}\in X\) if \(x^*\mathbf{a}=x^*\mathbf{b}\) for every \(x^*\in X^*\), then \(\mathbf{a}=\mathbf{b}\). Note that the dual of a locally convex topological vector space X separates points on X (see e.g. [12], Chapter 3: Corollary to Theorem 3.4; consult also Exercise 5(d) in the same chapter).

Corollary 3.10

Assume that \(f:[0,\infty )\rightarrow X\) is non-zero, for every \(x^*\in X^*\) the function \(x^*\circ f\) is right continuous at some point and \(g:[0,\infty )\rightarrow [0,\infty )\).

Then (fg) solves (7) if and only if

  1. 1.

    either

    $$\begin{aligned} f(s)=\mathbf{a}s,\quad s\in [0,\infty ),\qquad \text{ and }\qquad g\equiv 1, \end{aligned}$$
    (18)

    where \({\mathbf{a}}\in X\setminus \{\mathbf{0}\}\),

  2. 2.

    or

    $$\begin{aligned} f(s)=\mathbf{a}\log (\alpha s+1)\quad \text{ and }\quad g(s)=\alpha s+1,\quad s\in [0,\infty ), \end{aligned}$$
    (19)

    where \(\alpha >0\) and \({\mathbf{a}}\in X\setminus \{\mathbf{0}\}\).

Proof

Note that for any \(x^*\in X^*\) the pair \((x^*\circ f,\,g)\) solves (7). Moreover, since f is non-zero, \(x^*\circ f\) is non-zero for some \(x^*\in X^*\) and it follows from Theorem 3.3 that

$$\begin{aligned} g(s)=\alpha s+1,\quad s\in [0,\infty ), \end{aligned}$$

where \(\alpha \ge 0\).

If \(\alpha =0\), then f is additive and for every \(x^*\in X^*\) the additive and right continuous at a point function \(x^*\circ f\) has the form

$$\begin{aligned} x^*f(s)=(x^*f(1))\,s,\quad s\in [0,\infty ), \end{aligned}$$

whence \(f(s)=f(1)s\) for \(s\in [0,\infty )\), and we have (18) with \({\mathbf{a}}=f(1)\in X\setminus \{\mathbf{0}\}\).

If \(\alpha >0\), then by Theorem 3.3 for every \(x^*\in X^*\) either \(x^*\circ f\equiv 0\), or

$$\begin{aligned} x^*f(s)=a\log (\alpha s+1),\quad s\in [0,\infty ), \end{aligned}$$

where \(0\ne a\in {\mathbb R}\); in the second case

$$\begin{aligned} a=x^*f\left( \tfrac{e-1}{\alpha }\right) . \end{aligned}$$

Consequently, for every \(x^*\in X^*\) in both cases we have

$$\begin{aligned} x^*f(s)=x^*f\left( \tfrac{e-1}{\alpha }\right) \,\log (\alpha s+1),\quad s\in [0,\infty ), \end{aligned}$$

i.e.

$$\begin{aligned} f(s)=f\left( \tfrac{e-1}{\alpha }\right) \,\log (\alpha s+1),\quad s\in [0,\infty ). \end{aligned}$$

Thus we get (19) with \({\mathbf{a}}=f\left( \tfrac{e-1}{\alpha }\right) \in X\setminus \{\mathbf{0}\}.\)

\(\square \)

4 CGS Equations on a Vector Space

Throughout this section \(U=V\), a vector space over a field \(\mathbb F\). As it has already been observed, see Remark 2.3, if (fh) solves (6) on V, then \(\mathrm {Ker}(h)=\emptyset \). Therefore, due to Remark 2.2, equations (6) and (7) are equivalent with \(hg\equiv 1\).

We assume that \((M,+)\) is a unital magma with the two-sided cancelation property.

Theorem 4.1

Assume that \(f:V\rightarrow M\) is a non-zero function and \(g:V\rightarrow \mathbb F\).

Then (fg) solves (7) if and only if f is additive and \(g\equiv 1\).

Proof

First, we prove that f is additive, i.e.

$$\begin{aligned} f(s+t)=f(s)+f(t) \end{aligned}$$
(20)

for all \(s,t\in V\).

To this end we observe that for \(s\in V\), \(f(s)\ne \mathbf{0}\) implies \(g(s)=1\). Indeed, if \(g(s)\ne 1\), then \(s+g(s)t=t\) for \(t=\tfrac{s}{1-g(s)}\). Thus (7) for such s and t yields \(f(s)=\mathbf{0}\), a contradiction. Consequently, (20) holds true for \(s,t\in V\) such that at least one of f(s) and f(t) is not zero. To see this fact consider separately the cases: (a) both f(s) and f(t) are non-zero, (b) exactly one of f(s) and f(t) is non-zero. In the latter case use the identity \(\mathbf{a}+\mathbf{0}=\mathbf{0}+\mathbf{a}=\mathbf{a}\), which holds for every \(\mathbf{a}\in M\).

It suffices to prove (20) for \(s,t\in V\) such that \(f(s)=f(t)=\mathbf{0}\). Assume (20) does not hold for such \(s,t\in V\), i.e. \(f(s+t)\ne \mathbf{0}\). Then \(g(s+t)=1\) and (7) implies

$$\begin{aligned} f(t)=f(s+t-s)=f(s+t)+f(-s). \end{aligned}$$
(21)

Note that \(f(-s)=\mathbf{0}\). Otherwise \(g(-s)=1\) and (7) yields \(\mathbf{0}=f(-s+s)=f(-s)+f(s)\), a contradiction. Therefore, by (21) we get \(f(t)=f(s+t)\ne \mathbf{0}\), a contradiction.

Second, we prove that \(g\equiv 1\). Assume not, i.e. \(g(s)\ne 1\) for some \(s\in V\). For arbitrary \(u\in V\) set \(t:=\tfrac{u}{g(s)-1}\in V\). Then, by (20), we have

$$\begin{aligned} f(u)=f((g(s)-1)t)=f(s+g(s)t-s-t)=f(s+g(s)t)+f(-s-t). \end{aligned}$$
(22)

But (7) and (20) yield \(f(s+g(s)t)=f(s)+f(t)=f(s+t)\) and \(f(s+t)+f(-(s+t))=\mathbf{0}\). Consequently, (22) implies \(f\equiv \mathbf{0}\), a contradiction. \(\square \)