 Research
 Open Access
 Published:
Computation of Green’s functions through algebraic decomposition of operators
Boundary Value Problems volume 2016, Article number: 167 (2016)
Abstract
In this article, we use linear algebra to improve the computational time for obtaining Green’s functions of linear differential equations with reflection (DER). This is achieved by decomposing both the ‘reduced’ equation (the ODE associated with a given DER) and the corresponding twopoint boundary conditions.
Introduction
Differential operators with reflection have recently been of great interest, partly due to their applications to supersymmetric quantum mechanics [1–3] or topological methods applied to nonlinear analysis [4].
In the last years, the works in this field have been related to either obtaining eigenvalues and explicit solutions of different problems [5–8], their qualitative properties [4, 9], or obtaining the associated Green’s function [10–16]. In [16], the authors described a method to derive the Green’s function of a differential equation with constant coefficients, reflection and twopoint boundary conditions. This algorithm was implemented in Mathematica (see [17]) in order to put it to a practical use. Unfortunately, it was soon observed that, although theoretically correct, there were severe limitations when computing the Green’s functions of problems of high order. In this respect, we have to point out that an nthorder linear DER is reduced to a (2n)thorder ordinary differential equation; see Theorem 2.5 and compare equations (2.4) and (2.5). This particularity posses a great challenge since the computational time increases greatly with n.
To sort this out, the best option is to go back from a \((2n)\)thorder problem to two problems of order n. This procedure, compared to solving directly the order 2n, is much faster. Furthermore, it also happens that, in some cases, the decomposition provides two equivalent problems or a problem and its adjoint. In those cases, the improvement is even more notorious.
In the next section, we contextualize the problem with a brief introduction to differential equations with reflection and state some basic results concerning the Green’s function associated with them. In Section 3, we develop some theoretical results, which provide a way of decomposing the DER we are dealing with. Finally, in Section 4, we establish a suitable decomposition for the boundary conditions, state criteria for selfadjointness of the decomposed problem, and provide examples to illustrate the theory.
Differential equations with reflection
In order to establish a useful framework to work with these equations, we consider the differential operator D, the pullback operator of the reflection \(\varphi(t)=t\), denoted by \(\varphi^{*}(u)(t)=u(t)\), and the identity operator Id.
Let \(T\in{\mathbb{R}}^{+}\) and \(I:=[T,T]\). We now consider the algebra \({\mathbb{R}}[D,\varphi^{*} ]\) consisting of the linear operators of the form
where \(n\in{\mathbb{N}}\), \(a_{k},b_{k}\in{\mathbb{R}}\), \(k=1,\ldots,n\), which act as
on any function \(u\in W^{n,1}(I)\). The operation in the algebra is the usual composition of operators; we will omit the composition sign. We observe that \(D^{k}\varphi ^{*}=(1)^{k}\varphi^{*}D^{k}\) for \(k=0,1,\ldots\) , which makes it a noncommutative algebra. We will consider, for convenience, the sums \(\sum_{k=0}^{n}\equiv\sum_{k}\) such that \(k\in\{0,1,\ldots\}\), but taking into account that the coefficients \(a_{k}\), \(b_{k}\) are zero for big enough indices.
Notice that \({\mathbb{R}}[D,\varphi^{*}]\) is not a unique factorization domain. For instance,
Let \({\mathbb{R}}[D]\) be the ring of polynomials with real coefficients on the variable D. The following property is crucial for the obtaining of a Green’s function.
Theorem 2.1
([16], Theorem 2.1)
Take L be as defined in (2.1) and define
Then \(RL=LR\in{\mathbb{R}}[D]\).
Remark 2.2
If \(S:=RL=\sum_{k=0}^{2n} c_{k}D^{k}\), then
This implies that the reduced operator RL has only coefficients for the even powers of the derivative, so the equation is selfadjoint. If the boundary conditions are appropriate (we will clarify this statement in Theorem 4.4), then the Green’s function is symmetric [18]. Observe that \(c_{0}=a_{0}^{2}b_{0}^{2}\). Also, if \(L=\sum_{i=0}^{n} (a_{i}\varphi^{*}+b_{i} )D^{i}\) with \(a_{n}\ne0\) or \(b_{n}\ne0\), then we have that \(c_{2n}=(1)^{n}(a_{n}^{2}b_{n}^{2})\). Hence, if \(a_{n}=\pm b_{n}\), then \(c_{2n}=0\). This shows that composing two elements of \({\mathbb{R}}[D,\varphi^{*} ]\), we can get another element that has simpler terms in the sense of derivatives of lower order. This is quite difficult when computing the Green’s functions since, in this case, we could have one, many, or no solutions of our original problem [16]. The following example is quite illustrative.
Example 2.3
Consider the equation
This equation cannot have a solution since the lefthand side is an even function whereas the righthand side is an odd function.
As we said before, \(S=RL\) is a usual differential operator with constant coefficients. Consider now the following problem:
The existence of Green’s functions for problems such as (2.3) is a classical result (see, e.g., [19]). Here we present it adapted to our framework.
Theorem 2.4
Assume that the following homogeneous problem has a unique solution:
Then there exists a unique function, called Green’s function, such that:

(G1)
G is defined on the square \(I^{2}\).

(G2)
The partial derivatives \(\frac{\partial^{k}G}{\partial t^{k}}\) exist and are continuous on \(I^{2}\) for \(k=0,\ldots,n2\).

(G3)
\(\frac{\partial^{n1}G}{\partial t^{n1}}\) and \(\frac {\partial^{n}G}{\partial t^{n}}\) exist and are continuous on \(I^{2}\backslash\{(t,t) : t\in I\}\).

(G4)
The lateral limits \(\frac{\partial^{n1}G}{\partial t^{n1}}(t,t^{+})\) and \(\frac{\partial^{n1}G}{\partial t^{n1}}(t,t^{})\) exist for every \(t\in(a,b)\), and
$$ \frac{\partial^{n1}G}{\partial t^{n1}}\bigl(t,t^{}\bigr)\frac{\partial^{n1}G}{\partial t^{n1}}\bigl(t,t^{+}\bigr)= \frac {1}{a_{n}}. $$ 
(G5)
For each \(s\in(a,b)\), the function \(G(\cdot,s)\) is a solution of the differential equation \(Su=0\) on \(I\backslash\{s\}\).

(G6)
For each \(s\in(a,b)\),the function \(G(\cdot,s)\) satisfies the boundary conditions \(B_{k}u=0\), \(k=1,\ldots,n\).
Furthermore, the function \(u(t):=\int_{a}^{b}G(t,s)h(s)\,\mathrm{d}s\) is the unique solution of problem (2.3).
Now we can state the result that relates the Green’s function of a problem with reflection to the Green’s function of its associated reduced problem.
In order to do that, given an operator \({\mathscr {L}}\) defined on some set of functions of one variable, we will define the operator \({\mathscr {L}}_{\vdash}\) as \({\mathscr {L}}_{\vdash}G(t,s):={\mathscr {L}}(G(\cdot,s))_{t}\) for every s and any suitable function G of two variables.
Theorem 2.5
([16])
Let \(I=[T,T]\). Consider the problem
where L is defined as in (2.1), \(h\in L^{1}(I)\), and
Then, there exists \(R\in{\mathbb{R}}[D,\varphi^{*} ]\) (as in (2.2)) such that \(S:=RL\in{\mathbb{R}}[D]\), and the unique solution of problem (2.4) is given by \(\int_{a}^{b}R_{\vdash}G(t,s)h(s)\,\mathrm{d}s\), where G is the Green’s function associated with the problem
assuming that it has a unique solution.
As stated in Section 1, Theorem 2.5 was implemented in Mathematica in [17]. We now proceed to describe some steps that could be added to the algorithm in order to improve it.
Decomposing the reduced equation
The computation of Green’s functions is prohibitive in computation time terms [17], mostly for highorder equations, so it is necessary to find ways to palliate this problem. Our approach consists of decomposing our problem in order to deal with equations of lower order.
First, observe that from Remark 2.2 we know that the reduced equation has no derivatives of odd indices. For convenience, if p is a real (complex) polynomial, then we denote by \(p_{}\) the polynomial with the same principal coefficient and opposite roots.
Lemma 3.1
Let \(n\in{\mathbb{N}}\), and let \(p(x)=\sum_{k=0}^{n}\alpha _{2k}x^{2k}\) be a real polynomial of order 2n. Then there is a complex polynomial q of order n such that \(p=\alpha_{2n}qq_{}\). Furthermore, if \(\tilde{p}(x)=\sum_{k=0}^{n}\alpha_{2k}x^{k}\) has no negative roots, then q is a real polynomial.
Proof
First, observe that p is a polynomial on \(x^{2}\), and therefore, if λ is a root of p, so is −λ. Hence, using the fundamental theorem of algebra, the first part of the result can be derived by separating the monomials that compose p in two different polynomials with opposite roots.
Let us explicitly show that in the case p̃ has no negative roots, q is a real polynomial.
Take the change of variables \(y=x^{2}\). Then, \(p(x)=\tilde{p}(y)\), and, by the fundamental theorem of algebra,
for some integers σ, m, m̅, l and real numbers \(\lambda _{1},\ldots,\lambda_{\overline{m}}\), \(\nu_{1},\ldots,\nu_{l}\), \(\mu_{1},\ldots,\mu _{l}\) such that \(\lambda_{k}>0\) and \(\nu_{k}>\mu_{k}/2\) for every k in the appropriate set of indices. The terms of the form \(y^{2}+\mu_{k}y+\nu _{k}^{2}\) correspond to the pairs of complex roots of the polynomial. This means that the discriminant \(\Delta=\mu_{k}^{2}4\nu_{k}<0\), that is, \(\nu _{k}>\mu_{k}/2\).
Hence,
Now we have
for any k in the appropriate set of indices. Define
and
We have that \(p=\alpha_{2n}qq_{}\).
Observe that if λ is a root of p, then \(\lambda^{2}\) is a root of p̃. Hence, if p̃ has no negative roots, then p has no roots of the form \(\lambda=ai\) with \(a\ne0\). Thus,
that is, q is a real polynomial. □
Remark 3.2
Descartes’ rule of signs establishes that the number of positive roots (with multiple roots counted separately) of a real polynomial on one variable either is equal to the number of sign differences between consecutive nonzero coefficients or is less than it by an even number, provided that the terms of the polynomial are ordered by descending variable exponent. This implies that for a polynomial \(p(x)\) to have no negative roots, it suffices that all coefficients of \(p(x)\) are positive, that is, \(p(x)\) has positive even coefficients and negative odd coefficients.
There exist algorithmic ways of determining the exact number of positive (or real) roots of a polynomial. For more information on this issue, see, for instance, [20–22].
The following lemma establishes a relation between the coefficients of q and \(q_{}\).
Lemma 3.3
Let \(n\in{\mathbb{N}}\), and let \(q(x)=\sum_{k=0}^{n}\alpha_{k}x^{k}\) be a complex polynomial. Then
Proof
We proceed by induction. For \(n=1\), \(q(x)=\alpha(x\lambda_{1})\). Clearly, q has the root \(\lambda_{1}\), and \(q_{}(x)=\alpha(x+\lambda _{1})=(1)^{1+1}\alpha x+(1)^{1}\alpha\lambda_{1}\) has the root \(\lambda_{1}\).
Assume that the result is true for some \(n\ge1\). Then, for \(n+1\), q is of the form \(q(x)=(x\lambda_{n+1})r(x)\), where \(r(x)=\sum_{k=0}^{n}\alpha_{k}x^{k}\) is a polynomial of order n, that is,
Now, \(q_{}(x)=(x+\lambda_{n+1})r_{}(x)\). Since the formula is valid for n,
So the formula is valid for \(n+1\) as well. □
Remark 3.4
The result can be directly proven by considering the last statement in Remark 3.2. If we take a polynomial \(p(x)=a(x\lambda_{1})\cdots(x\lambda_{n})\), then the polynomial \(p(x)\) has exactly opposite roots. In fact, \(p(x)=a(x\lambda _{1})\cdots(x\lambda_{n})=(1)^{n}a(x+\lambda_{1})\cdots(x+\lambda_{n})\). It is easy to check that the coefficients of \(p(x)\) are precisely as described in Lemma 3.3 save for the factor \((1)^{n}\).
This last lemma allows the computation of the polynomials q and \(q_{}\) related to the polynomial RL on the variable D using the formula given in Remark 2.2. We will assume that RL is of order 2n, that is, \(a_{n}^{2}b_{n}^{2}\neq0\). Otherwise, the problem of computing q and \(q_{}\) would be the same, but these polynomials would be of lower order. Also, assume that RL, considered as a polynomial on \(D^{2}\), has no negative roots in order for q to be a real polynomial. If \(L=\sum_{k=0}^{n}(a_{k}\varphi^{*}+b_{k})D^{k}\) and \(q(D)=D^{n}+\sum_{k=0}^{n1}\alpha_{k}D^{k}\), then
This relation establishes the following system of quadratic equations:
where \(a_{k},b_{k},\alpha_{k}=0\) if \(k\notin\{0,\ldots,n\}\) and \(\alpha _{n}=1\). These are n equations with n unknowns \(\alpha_{0},\ldots ,\alpha_{n}\). We present here the case of \(n=2\) to illustrate the solution of these equations.
Example 3.5
For \(n=2\), we have that
and the system of equations is
Before computing the solutions, let us state explicitly the limitations that RL, considered as an order 2 polynomial on \(D^{2}\), that is, \(RL(x)=a x^{2}+b x +c\), has no negative roots implies. There are two options:

(I)
There are two complex roots, that is, \(\Delta= b^{2}4ac<0\). This is equivalent to \(ac>0\landb<2\sqrt{ac}\) or, expressed in terms of the coefficients of RL,
$$ \bigl(b_{0}^{2}a_{0}^{2}\bigr) \bigl(b_{2}^{2}a_{2}^{2}\bigr)>0 \quad \text{and}\quad \bigl\vert a_{1}^{2}+2 a_{0} a_{2}+b_{1}^{2}2 b_{0} b_{2} \bigr\vert < 2\sqrt{\bigl(b_{0}^{2}a_{0}^{2} \bigr) \bigl(b_{2}^{2}a_{2}^{2}\bigr)}. $$ 
(II)
There are two nonnegative roots, that is, \(\Delta =b^{2}4ac\ge0\), and
$$ \bigl(b+\sqrt{b^{2}4ac}\bigr)/(2a)\le0. $$This is equivalent to \((a,c\ge0\landb\ge2\sqrt{ac})\lor(a,c\le 0\land b\ge2\sqrt{ac})\) or, expressed in terms of the coefficients of RL,
$$ \Bigl[\bigl(b_{0}^{2}a_{0}^{2}\bigr), \bigl(b_{2}^{2}a_{2}^{2}\bigr)\ge0\land \bigl(a_{1}^{2}+2 a_{0} a_{2}+b_{1}^{2}2 b_{0} b_{2}\bigr)\ge2\sqrt{\bigl(b_{0}^{2}a_{0}^{2} \bigr) \bigl(b_{2}^{2}a_{2}^{2}\bigr)} \Bigr] $$or
$$ \Bigl[\bigl(b_{0}^{2}a_{0}^{2}\bigr), \bigl(b_{2}^{2}a_{2}^{2}\bigr)\le0\land \bigl(a_{1}^{2}+2 a_{0} a_{2}+b_{1}^{2}2 b_{0} b_{2}\bigr)\ge2\sqrt{\bigl(b_{0}^{2}a_{0}^{2} \bigr) \bigl(b_{2}^{2}a_{2}^{2}\bigr)} \Bigr]. $$
Now, with these conditions, the solutions of the system of equations (3.1) are as follows.
Case (I). We have two solutions:
Case (II). We have four solutions depending on whether we choose \(\xi=1\) or \(\xi=1\):
These solutions provide welldefined real numbers by conditions (I) and (II).
Decomposing the boundary conditions
Now we consider the cases where the problem can be decomposed into two equations. We will try to identify those circumstances when problem (2.5)(2.6)(2.7) can be expressed as an equivalent factored problem of the form
where \(S=L_{2}L_{1}\). If that where the case, then conditions (2.6)(2.7) would be equivalent to
In this case, the Green’s function of problem (2.5)(2.6)(2.7) can be expressed as
where \(G_{1}\) is the Green’s function associated with problem (4.1), and \(G_{2}\) is the one associated with problem (4.2), assuming that both Green’s functions exist.
In order to guarantee that (2.6)(2.7) and (4.3) are equivalent, let us establish the following definitions. Let
Then the boundary conditions (2.6) can be expressed as \(\Gamma_{1}\overline{X}_{n}+\Theta_{1}X_{n}=0\). In the same way, (2.7) can be written as \((\Gamma_{2}\ \Gamma_{3})\overline{X}_{2n}+(\Theta_{2}\ \Theta_{3})X_{2n}=0\) for some matrices \(\Gamma _{2},\Gamma_{3},\Theta_{2},\Theta_{3}\in{ \mathscr {M}}_{n}({\mathbb{R}})\). So, globally, the conditions on equation (2.5) can be expressed as
Now, assume that \(L_{1}\) and \(\widetilde{V}_{j}\) are of the form
for some \(c_{l},\gamma_{jk},\delta_{jk}\in{\mathbb{R}}\), \(l,j,k=1,\ldots ,n\), where \(a^{*}\) denotes the pullback by the constant a. Define now \(\Phi:=(\gamma_{jk})_{j,k}, \Psi:=(\delta_{jk})_{j,k}\in {\mathscr {M}}_{n}({\mathbb{R}})\), and
where \(\Xi_{1}, \Xi_{2}\in{ \mathscr {M}}_{n}({\mathbb{R}})\), \(\Xi_{2}\) is invertible (because \(c_{n}\ne0\)), and \(\Xi_{1}\) is invertible if and only if \(c_{0}\ne0\).
Now we are ready to start the calculations. We have that
Hence, we can write (4.3) in the form
Clearly, it is convenient to take \(\widetilde{\Phi}=\Gamma_{1}\) and \(\widetilde{\Psi}=\Theta_{1}\), that is, \(V_{j}=B_{j}\), \(j=1,\ldots,n\).
Lemma 4.1
If \(\Gamma_{1}\) and \(\Gamma_{3}\) are invertible and \(\Theta_{2}=\Gamma _{2}\Gamma_{1}^{1}\Theta_{1}+\Theta_{3}\Xi_{2}^{1}\Xi_{1}\Gamma_{3}\Xi _{2}^{1}\Xi_{1}\Gamma_{1}^{1}\Theta_{1}\), then, taking
condition (4.4) is equivalent to condition (4.5), and, therefore, problems (2.5)(2.6)(2.7) and (4.1)(4.2) are equivalent.
Proof
Let
The matrix A is invertible, and
Hence, conditions (4.4) and (4.5) are equivalent. □
Analogously, we have a result when \(\Theta_{1}\) and \(\Theta_{3}\) are invertible.
Lemma 4.2
If \(\Theta_{1}\) and \(\Theta_{3}\) are invertible and \(\Gamma_{2}=\Theta _{2}\Theta_{1}^{1}\Gamma_{1}+\Gamma_{3}\Xi_{2}^{1}\Xi_{1}\Theta_{3}\Xi _{2}^{1}\Xi_{1}\Theta_{1}^{1}\Gamma_{1}\), then, taking
condition (4.4) is equivalent to condition (4.5), and, therefore, problems (2.5)(2.6)(2.7) and (4.1)(4.2) are equivalent.
The following example illustrates this discussion explicitly.
Example 4.3
Consider the following problem:
where \(h(t)=\sin t\). Then, the operator we are studying is \(L=D^{3}+\varphi^{*}+1\). If we take \(R:=D^{3}+\varphi^{*}1\), then we have that \(RL=D^{6}\), which admits a simple decomposition in \({\mathbb{R}}[D]\) as \(RL=(D^{3})(D^{3})=L_{2}L_{1}\).
The boundary conditions are
Taking this into account, we add the conditions
Then our new reduced problem, writing the boundary conditions in matrix form, is
where \(f(t)=R h(t)=h'''(t)+h(t)h(t)=3\sin t\).
Now, we can check that we are working under the conditions of Lemma 4.1. We have that \(\Gamma_{1}=\Gamma_{3}=\mathrm{Id}\), \(\Gamma_{2}=\Theta_{2}=0\), and
On the other hand,
Thus, it is straightforward to check that
and therefore the hypotheses of Lemma 4.1 are satisfied. The conditions \(\widetilde{V}_{j}\) are given by the matrices \(\Phi=\mathrm{Id}\) and \(\Psi=\Xi_{2}\Gamma_{3}^{1}\Theta_{3}\Xi_{2}^{1}=\Theta_{3}\). Hence, we know that this problem is equivalent to the factored system
Thus, it is clear that
where \(G_{1}=G_{2}\) are, respectively, the Green’s functions of (4.8) and (4.9). The Green’s functions of problems involving linear ordinary differential equations with constant coefficients and twopoint boundary conditions can be computed with the Mathematica notebooks [23] or [17]. Explicitly,
Hence, the Green’s function G for problem (4.7) is given by
Therefore, using Theorem 2.5, the Green’s function for problem (4.6) is
Hence, the solution of problem (4.6) is given by
Computationally, this procedure poses a big advantage: it is always easier to obtain the Green’s function for two nthorder problems than for one (2n)thorder problem. Furthermore, if the hypotheses of Lemma 3.1 are satisfied and we are able to obtain a factorization of the aforementioned kind using q and \(q_{}\) in the place of \(L_{1}\) and \(L_{2}\), then we have an extra advantage: the differential equation given by \(q_{}\) is the adjoint equation of that given by q multiplied by the factor \((1)^{n}\). This fact, together with the following result (which can be found, although not stated as in this work, in [18]), illustrates that in this case it may be possible to solve problem (2.4) just computing the Green’s function of one nthorder problem.
Theorem 4.4
Consider an interval \(J=[a,b]\subset{\mathbb{R}}\), functions \(\sigma ,a_{i}\in\operatorname{L^{1}}(J)\), \(i=1,\ldots,n\), real numbers \(\alpha _{ij}\), \(\beta_{ij}\), \(h_{i}\), \(i=1,\ldots, n\), \(j=0,\ldots,n1\), a vector subspace \(D(L_{n})\subset W^{n,1}(J)\), the operator
with \(a_{0}=1\), and the problem
where
Then, the associated adjoint problem is
where
Furthermore, if \(G(t,s)\) is the Green’s function of problem (4.10), then that associated with problem (4.11) is \(G(s,t)\).
Hence, if we can decompose problem (2.5)(2.6)(2.7) in two adjoint problems of the form (4.1)(4.2), then its Green’s function is
where \(G_{1}\) is the Green’s function of (4.1), and \(G_{2}(t,s)=G_{1}(s,t)\) is that of (4.2). We note, though, that unless the operator \(q_{}\) is the adjoint equation times \((1)^{n}\), the boundary conditions may not be the adjoint ones.
Example 4.5
Consider the problem
Taking \(R=\varphi^{*}D+\sqrt{2}\varphi^{*}\mathrm{Id}\) and composing problem (4.12) with this operator, we obtain the reduced problem
Problem (4.13) is equivalent to the factored system
for \(t\in[1,1]\). Observe that problem (4.15) is the adjoint problem of (4.14). Since the Green’s function of problem (4.14) is given by
and, therefore, \(G_{1}(s,t)\) is the Green’s function of problem (4.15), the Green’s function of problem (4.13) is
Finally, the Green’s function of problem (4.12) is
Hence, the solution of problem (4.12) is
References
 1.
Post, S, Vinet, L, Zhedanov, A: Supersymmetric quantum mechanics with reflections. J. Phys. A, Math. Theor. 44(43), 435301 (2011)
 2.
Roychoudhury, R, Roy, B, Dube, PP: NonHermitian oscillator and Rdeformed Heisenberg algebra. J. Math. Phys. 54(1), 012104 (2013)
 3.
Gamboa, J, Plyushchay, M, Zanelli, J: Three aspects of bosonized supersymmetry and linear differential field equation with reflection. Nucl. Phys. B 543(1), 447465 (1999)
 4.
Cabada, A, Infante, G, Tojo, FAF: Nontrivial solutions of Hammerstein integral equations with reflections. Bound. Value Probl. 2013, 86 (2013)
 5.
Piao, D, Sun, J: Besicovitch almost periodic solutions for a class of second order differential equations involving reflection of the argument. Electron. J. Qual. Theory Differ. Equ. 2014, 41 (2014)
 6.
Piao, D, Xin, N: Bounded and almost periodic solutions for second order differential equation involving reflection of the argument (2013). arXiv:1302.0616
 7.
Kritskov, L, Sarsenbi, A: Spectral properties of a nonlocal problem for a secondorder differential equation with an involution. Differ. Equ. 51(8), 984990 (2015)
 8.
Kritskov, LV, Sarsenbi, AM: Basicity in \(L_{p}\) of root functions for differential equations with involution. Electron. J. Differ. Equ. 2015, 278 (2015)
 9.
Ashyralyev, A, Sarsenbi, AM: Wellposedness of an elliptic equation with involution. Electron. J. Differ. Equ. 2015, 284 (2015)
 10.
Sarsenbi, A: The Green’s function of the second order differential operator with an involution and its application. AIP Conf. Proc. 1676, 020010 (2015)
 11.
Sarsenbi, AA: Green’s function of the secondorder differential operator with involution from boundary conditions of Neumann. AIP Conf. Proc. 1676, 020074 (2015)
 12.
Cabada, A, Tojo, FAF: Comparison results for first order linear operators with reflection and periodic boundary value conditions. Nonlinear Anal. 78, 3246 (2013)
 13.
Cabada, A, Tojo, FAF: Solutions of the first order linear equation with reflection and general linear conditions. Glob. J. Math. Sci. 2(1), 18 (2013)
 14.
Cabada, A, Tojo, FAF: Existence results for a linear equation with reflection, nonconstant coefficient and periodic boundary conditions. J. Math. Anal. Appl. 412(1), 529546 (2014)
 15.
Cabada, A, Tojo, FAF: Solutions and Green’s function of the first order linear equation with reflection and initial conditions. Bound. Value Probl. 2014, 99 (2014)
 16.
Cabada, A, Tojo, FAF: Green’s functions for reducible functional differential equations. Bull. Malays. Math. Sci. Soc., 122 (2016)
 17.
Tojo, FAF, Cabada, A, Cid, JA, MáquezVillamarín, B: Green’s functions with reflection. http://library.wolfram.com/infocenter/MathSource/9087 (2014)
 18.
Cabada, A: Green’s functions in the theory of ordinary differential equations. Springer, Berlin (2014)
 19.
Cabada, A, Cid, JÁ: On the sign of the Green’s function associated to Hill’s equation with an indefinite potential. Appl. Math. Comput. 205(1), 303308 (2008)
 20.
Yang, L, Xia, B: Explicit criterion to determine the number of positive roots of a polynomial. MM Res. Prepr. 15, 134145 (1997)
 21.
Yang, L, Hou, XR, Zeng, ZB: A complete discrimination system for polynomials. Sci. China Ser. E 39(6), 628646 (1996)
 22.
Liang, S, Zhang, J: A complete discrimination system for polynomials with complex coefficients and its automatic generation. Sci. China Ser. E 42(2), 113128 (1999)
 23.
Cabada, A, Cid, JA, MáquezVillamarín, B: Green’s functions computation. http://library.wolfram.com/infocenter/MathSource/8825 (2014)
Acknowledgements
The author wants to acknowledge his gratitude to the anonymous referee for helping improve the manuscript, especially in the proof of Lemma 3.1. This work was partially supported by Consellería de Cultura, Educación e Ordenación Universitaria, Xunta de Galicia, Spain, project EM2014/032 and supported by FPU scholarship, Ministerio de Educación, Cultura y Deporte, Spain.
Author information
Additional information
Competing interests
The author declares that they have no competing interests.
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.
About this article
Cite this article
Tojo, F.A.F. Computation of Green’s functions through algebraic decomposition of operators. Bound Value Probl 2016, 167 (2016) doi:10.1186/s136610160671y
Received:
Accepted:
Published:
MSC
 34Bxx
 47Lxx
 34Kxx
Keywords
 Green’s functions
 ODE
 reflection
 decomposition