Generic selectors
Exact matches only
Search in title
Search in content
Post Type Selectors
Search in posts
Search in pages
Filter by Categories
Editorial
Media & News
Original Article
Review Article
Generic selectors
Exact matches only
Search in title
Search in content
Post Type Selectors
Search in posts
Search in pages
Filter by Categories
Editorial
Media & News
Original Article
Review Article
View/Download PDF

Translate this page into:

Original Article
2 (
1
); 3-11
doi:
10.25259/STN_35_2025

Development of Mamadu–Njoseh Polynomial Neural Networks for Physics-Informed Learning

Department of Mathematics, Delta State University, Abraka, Nigeria
Department of Mathematics, University of Delta, Agbor, Delta State, Nigeria.
Author image

*Corresponding author: Dr. Ebimene James Mamadu, Department of Mathematics, Delta State University, Abraka, Nigeria. emamadu@delsu.edu.ng

Licence
This is an open-access article distributed under the terms of the Creative Commons Attribution-Non Commercial-Share Alike 4.0 License, which allows others to remix, transform, and build upon the work non-commercially, as long as the author is credited and the new creations are licensed under the identical terms.

How to cite this article: Mamadu EJ, Nwankwo JC, Njoseh NI. Development of Mamadu–Njoseh Polynomial Neural Networks for Physics-Informed Learning. Sci Tech Nex. 2026;2:3-13. doi: 10.25259/STN_35_2025

Abstract

Objective

This present work seeks to address the challenges in solving differential equations through the development of an improved Physics-informed neural network approach aimed at ensuring efficient convergence and precise solutions.

Material and Methods

A novel approach, termed the Mamadu-Njoseh polynomial physics-informed neural Network (MNP-PINN), is introduced. This approach utilises the orthogonal Mamadu-Njoseh polynomials (MNP) as activation functions in the network architecture. Physical constraints arising from the governing differential equation are embedded into the loss function to guide the learning process. The performance validation for the new approach is carried out by testing it with the one-dimensional Poisson equation. All computations are implemented in Python 3.11, with consistent procedures for network initialisation, collocation point generation, and training to ensure reproducibility.

Results

The results show excellent convergence of MNP-PINN solutions with the analytical solution of the Poisson equation. The solution has spectral accuracy with errors consistently small throughout the solution domain. In addition, there is stability in the optimisation process of the training algorithm, showing improved efficiency in the machine learning process compared with the analytical PINN solution.

Conclusion

The MNP-PINN has provided a robust and highly accurate framework for the solution of the differential equations. By utilizing the concepts of orthogonality and approximation provided by the MNP, improvement has been made in convergence acceleration and accuracy to that provided in PINNs. It appears from the results that the MNP-PINN framework has tremendous potential for further generalisation and expansion to more complex models.

Keywords

Machine learning
Mamadu–Njoseh polynomials
Network activation function
Physics-informed neural networks
Poisson equation

1. INTRODUCTION

Scientific machine learning has experienced a paradigm shift in the past several years, driven by the integration of data-driven methods and physics-based modelling.[1-3] Among these advancements, Physics-Informed Neural Networks (PINNs) have been revolutionary in the solution of differential equations through the inclusion of physical constraints governing equations, boundary, and initial conditions, within the training loss function.[4,5] In contrast to traditional data-driven models that only rely on empirical data, our PINNs enforce compliance with the underlying physical laws, thus recovering generalisable, interpretable, and physically relevant solutions.[6,7]

However, despite their success, conventional PINNs still consistently suffer from limitations. Traditional activation functions, such as tanh, rectified linear unit (ReLu), and sigmoid, cannot reflect the mathematical or physical properties of many systems, particularly those with oscillatory, fractional, or stochastic dynamics.[8,9] They often lack orthogonality and weighted smoothness, thus, numerical inefficiencies, slow convergence, and deeper network architectures are necessitated.[10,11] Besides, they perform poorly when they are multi-scale, nonlocal, or memory-dependent processes involved, for which the classical neural representations fail to capture the underlying structure of the solution.[12]

Due to such constraints, there is a developing research direction to embed mathematical basis functions and orthogonal polynomial systems within neural architectures for enhancing approximation accuracy, interpretability, and convergence behaviour. Some recent works have successfully applied Chebyshev, Legendre, Hermite, and Jacobi polynomial bases in neural approximators to enhance spectral expressiveness and learning efficiency.[13,14,9] While these polynomial-based neural architectures have been promising, they are themselves restricted to specific weight functions and do not have flexibility in approximating nonlinear, stochastic, and fractional operators.

In order to address these challenges, the present study suggests a new class of physics-informed neural architectures based on the Mamadu–Njoseh Polynomials (MNP), a system of orthogonal polynomials defined on the interval [1,1] with the weight functionw(x)=1+x2 .[15] These polynomials possess desirable mathematical properties, including orthogonality, stability, and normalised boundary behaviour ( φn(1)=1 ), which make them ideal candidates for developing mathematically stable and physically consistent activation functions. Their use in neural architectures yields the Mamadu–Njoseh Polynomial Physics-Informed Neural Network (MNP-PINN), a hybrid model that fuses orthogonal polynomial theory and physics-informed deep learning.

The motivation for the current research is the need to develop mathematics-aware learning systems that exceed conventional data-fitting paradigms by providing structured and physically constrained approximations. The proposed MNP-PINN introduces MNP basis functions as the primitive computational elements, allowing the network to efficiently approximate both the solution and its derivatives while ensuring consistency with the underlying physical laws. Through this coupling, the MNP-PINN model aims to improve training stability, convergence rate, and accuracy of solutions, especially for systems that are modelled using fractional, stochastic, and wave-type equations.[9,13]

To this end, this study develops a rigorous theoretical and computational framework for the Mamadu–Njoseh Polynomial Physics-Informed Neural Network (MNP-PINN), grounded in orthogonal polynomial basis functions. The framework establishes foundational mathematical properties of the proposed architecture, including orthogonality, convergence, and numerical stability, under the weighted inner-product structure induced by the weight function w(x)=1+x2 . Furthermore, the governing physical laws are seamlessly incorporated into the MNP-PINN loss formulation to ensure physically consistent learning.

To validate the effectiveness of the proposed approach, the MNP-PINN is benchmarked against exact analytical solutions, with performance evaluated in terms of solution accuracy, computational efficiency, and numerical robustness. Numerical experiments demonstrate the ability of the MNP-PINN to efficiently approximate physically meaningful solutions, showcasing its potential for solving complex differential equations beyond traditional data-driven paradigms. In particular, the one-dimensional Poisson equation is employed as a model problem to assess the method’s performance. This equation, despite its apparent simplicity, plays a fundamental role in mathematical physics and computational science. It encapsulates key characteristics of elliptic partial differential equations, serving as a canonical benchmark for evaluating approximation theories, assessing stability and convergence behaviour, and verifying computational methodologies. In physical applications, the Poisson equation governs a broad spectrum of steady-state phenomena, including electrostatic potential distribution, heat conduction processes, fluid pressure variations, and mechanical deformation under equilibrium.

Hence, the paper attempts to put the Mamadu–Njoseh Polynomial Neural Network on a novel, efficient, and theoretically sound foundation as a paradigm for physics-informed learning. By combining the rigour of orthogonal polynomial approximation theory and the flexibility of deep neural architecture, the MNP-PINN paradigm has the potential to revolutionize solution techniques for physical systems with nonlinearity, fractionality, and stochasticity, a feat of great value to computational mathematics and applied artificial intelligence.

2. MATERIAL AND METHODS

2.1. Weighted space and orthogonality

Let the weight inner product and associated norm be defined as[16]

(2.1)
u,ww= 1 1 wxuxvxdx, u Lw2 = u,vw 2 ,

where wx is a prescribed weight function.

The MNP φn n0 form a sequence of orthogonal polynomials such that

(2.2)
φn= i=0 nai xi, ai0,

satisfying

(2.3)
φn,φm=0 for nm,

and the normalisation condition

(2.4)
φn 1 =1.

Using properties (2.2)–(2.4) with the weight wx=1+x2 , the first three MNP are obtained as

φ0 x=1,φ1 x=x, φ2 x= 1 3 5 x2 2 .

Similarly,

φ0 Lw2 2 = 8 3 , φ1 Lw2 2 = 16 15 , φ2 Lw2 2 = 136 189 ,

which were obtained by exact polynomial integration on 1,1 .

Analogue to classical weighted Sobolev spaces, for integer s0, define

Hws 1,1 = uLw2 :ukL uk 2 for 0ks ,

where wk x denotes an appropriate family of derivative weights.[17] The corresponding norm is

(2.5)
u Hws 2 = k=0 s u k Hws 2 = k=0 s 1 1 u k x 2 wxdx.

Because φn are orthogonal in Lw2 1,1 , every uLw2 1,1 has the expansion

ux= n=0 cn φn x,

where the spectral coefficients are given by

cn= u,φnw φn Lw2 2 .

To facilitate analysis, we introduce the spectral Sobolev norm[18]

(2.6)
u s,w 2 = n=0 1+λn s cn 2 φn Lw2 2 ,

where λn n0 is a sequence of positive numbers associated with the spectral scaling (typically related to eigenvalues of a Sturm–Liouville operator).

Let SN be the weighted orthogonal projection onto span φ0 ,,φN , that is,

(2.7)
SNu= n=0 N u,φnw φn Lw2 2 φn.

If uHws with integer s0, then for 0rs,

uSNu Hwr kNrs u Hws ,

where k>0 is a constant. This result establishes algebraic convergence of the projection error for functions of Sobolev regularity.

Similarly, if u is analytic in an ellipse of the complex plane that contains the interval [1,1], then the projection SNu achieves spectral convergence,[19]

uSNu Lw2 keαN ,

for some constant α>0 independent of N.

2.2. Mamadu-Njoseh polynomial neuron and layer

Let HLw2 1,1 denote the weighted Hilbert space defined by

(3.1)
H= u: 1,1 : u w2 = 1 1 1+x2 ux 2 dx< ,

equipped with the weighted inner product

(3.2)
u,vw= 1 1 1+x2 uxvxdx.

The MNP φn x n=0 form a complete orthogonal basis for H, satisfying

(3.3)
φn,φmw=γn δmn , δmn >0,

where δmn is the Kronecker delta, and γn is the normalisation constant associated with the nth basis function φn.

A neuron in a MNP layer computes

(3.4)
hj=φn Wjx+bj ,

where Wj 1×d and bj are trainable parameters for neuron j, and φn is the nth MNP basis function.

Thus, the hidden transformation

Φ:dm

is given[20] by

(3.5)
Φx= φn W1 x+b1 ,φn W2 x+b2 ,, φn Wmx+bm T,

which maps the input vector x into a nonlinear feature space spanned by the orthogonal polynomial basis φ0 ,,φn ,.

The operator Φ therefore enriches the representational capability of the network by embedding the input in a structured polynomial function space, ensuring controlled smoothness and improving numerical stability relative to generic activation functions.

2.3. MNP operator and polynomial neuron mapping

Each MNP φn x admits a finite polynomial expansion

(3.6)
φn x= k=0 nan,k xk,

where the coefficients an,k are determined by the weighted orthogonality condition

1 1 1+x2 φn x φm xdx=0, for all mn.

This ensures that φn x lies in the orthogonal complement of the polynomial subspace spanned by φ0 ,,φn1 .

Consequently, φn x can be constructed using properties (2.2)–(2.4) with the weight wx=1+x2 . Thus, each MNP basis element is generated by removing lower-order components, ensuring strict orthogonality in the weighted Hilbert space.

In the context of neural networks, each neuron of the form

zj=Wjx+bjhj =φn zj ,

nonlinearly lifts the input into a higher-order orthogonal polynomial feature space. This performs an implicit orthogonal projection onto the nth spectral mode, enriching the representation and enhancing numerical stability relative to generic activations.[20,21]

Define the MNP operator ρ acting on an affine input

z=Wx+b

as

(3.7)
ρnz x=φn zx = k=0 nan,k zx k.

Thus, a single MNP neuron computes the nonlinear mapping

(3.8)
hj=ρn zj= k=0 nan,k Wjx+bj k.

where Wj and bj ​ are the weight and bias for neuron j.

Applying the binomial expansion yields

(3.9)
hj= k=0 nan,k r=0 k n k Wjr bjkr xr

This representation shows that each MNP neuron induces a structured polynomial mapping of degree n in the input variables. In particular, the terms Wjx r contain mixed products of input features, meaning that nonlinear cross-feature interactions are introduced through the multinomial expansion of Wjx.

2.4. Differential properties and learning dynamics of an MNP layer

Consider a neural network (or a single neuron) employing the nth order MNP φn ​ as its activation function. For neuron j, define the affine transformation

zj x=Wjx+bj,

and the activation output

hj x=φn zj x ,

where the polynomial form of φn ​ is given by

(3.11)
φn u= k=0 nan,k uk,an,n 0.

Assuming the inputs xd , the network output can be represented as

(3.12)
F x;θ = j=1 mθj φn wjx+bj ,

where the collection of trainable parameters is

Θ= θj,wj,bj j=1 m,

and θj are output-layer weights. The model thus forms a polynomial neural network where each neuron expands the input into an n-degree orthogonal polynomial basis component.

Since the pre‐activation for neuron j is

zj=wjx+bj,

the first and second derivatives of the MNP activation follow directly from the polynomial definition,

(3.13)
φn u= k=1 nkan,k uk1 ,

(3.14)
φn ’’ u= k=1 nk k1 an,k uk2 .

These closed forms enable efficient Jacobian and Hessian computation, critical for physics-informed neural network (PINN) training where higher–order derivatives appear in the PDE residual.

Parameter gradients for neutron j can be computed as,

hj Wj =φn zj xT 1×d ,

hj bj =φn zj ℝ,

hj θj =φn zj .

Similarly, the gradient of the network output is given as,

uj Fx=θj φn zj xT 1×d , bjF=θj φn zj , θj F=φn zj .

2.5. Gradient magnitudes

Let R>0 be a bound on zj over the data distribution or during training. Define, for integer 0pn0 ,

(3.15)
An p R= k=pn k p an,k Rkp ,

then the pth derivative of φn ​ satisfies the uniform bound

max u R φn p u An p Rp!.

The first–derivative bound of activation is defined by

(3.16)
max u R φn u k=1 nk an,k Rk1 =Bn R.

This implies that if x 2 X, then the operator norm of the Jacobian w.r.t Wj satisfies

(3.17)
WjF 2 = θj φn zj x 2 θj Bn RX.

Thus, the local Lipschitz constant of neuron j with respect to x is bounded by θj Bn(R) , where Bn(R) controls derivative growth. Large Bn(R) induces large gradients and potential instability, while small Bn(R) may lead to vanishing gradients.

There are two primary mechanisms responsible for vanishing and exploding gradients in MNP–based networks:[1]

  • a.

    Since φn' u is a polynomial of degree (n1), the following scenario exists;

  • i.

    If z1 and coefficients are moderate, φn' z can grow as O z n1 , implying exploding gradients.

  • ii.

    If z1 and higher–order coefficient dominate but small, φn' zan,1 implying small gradients.

  • b.

    Gradients for Wj scale with θj implies the following possibilities;

  • i.

    Large θj amplify, and

  • ii.

    Regularising or initialising θj small mitigates an explosion.

A practical bound of the gradient to avoid explosion is to enforce[22]

(3.18)
k=1 nk an,k Rk1 G,

such that

φn z Gforall zR.

However, (3.18) can be satisfied by a scaling coefficient an,k or enforcing R small via weight normalisation.

2.6. MNP-PINN construction

Let Ωd be the spatial domain and I= 0,T the time interval. Consider a partial differential equation (PDE) operator acting on the scalar field u x,t defined as

(4.1)
[u] x,t =f x,t , x,t Ω×I,

with boundary conditions Bu=g on Ω×I and initial conditions u x,0 =u0 x, tu x,0 =u1 x. f x,t is the source term.

We seek an approximate uθ parametrized by θ that satisfies the PDE and boundary/initial conditions, in an informed sense.

Since φn ξ n=0 is orthogonal on 1,1 w.r.t. wξ=1+ξ2 and φn 1 =1 , then for any pre-activation ξ= x,t;θ we define a polynomial layer

N ξ= n=0 Nαn φn ξ,

with learned coefficients αn=αn θ.

The choice of network representation determines how the neural architecture captures and expresses the spatial-temporal dependencies in the solution uθ x,t . Two formulations are commonly considered.

  • a.

    Spectral form: The spectral form defines a global functional expansion of the neural approximation uθ x,t over a set of basis functions φn x,t n=0 N ​. It is expressed as

(4.2)
uθ x,t = n=0 Ncn x,t;θc φn x,t;θ ,

where cn are small neural networks that modulate each function.

  • b.

    Activation form: Alternatively, the activation form incorporates the orthogonal polynomial basis directly as an activation function within the hidden layers of the network. It is defined as

(4.3)
h l =N l W l h l1 +b l ,

such that the orthogonal polynomial structure is embedded in the layer.

In this paper, for analytical clarity and interpretability, the spectral form (4.2) is adopted. Under this formulation, the approximation is rewritten as:

(4.4)
uθ x,t = n=0 Ncn θ φn ξ x,t ,

where ξ is an affine map ξ=Bxx+Btt+b, or normalised coordinate mapping ξ=M x,t 1,1 . Bx,Bt,b are Learnable scaling and shift parameters used to normalise spatial and temporal coordinates into the standard domain [−1,1] [Figure 1].

Architecture of the proposed Mamadu-Njoseh polynomial physics informed neural network (MNP-PINN).
Figure 1:
Architecture of the proposed Mamadu-Njoseh polynomial physics informed neural network (MNP-PINN).

The MNP-PINN is implemented as a fully connected feedforward neural network. The network consists of one hidden layer with 40 neurons. The hidden layer employs quadratic Mamadu–Njoseh polynomials (MNPs) as activation functions, providing orthogonality and stability benefits over conventional activations. The output layer is linear to produce predicted solution values, as shown in Figure 1. The input dimension is d=1 for the one-dimensional Poisson problem.

2.7. Physics-informed loss

Let xi,ti i=0 Nr Ω×I be sampled collocation points for PDE residuals, boundary points xj,tj Ω×I, and initial points xk k=1 N0 Ω. The physics-informed loss is given as

(4.5)
Lθ=Lres +λb Lbnd +λ0 Linit +λreg Lreg ,

where,

Lres = 1 Nr i=1 Nr M uθ xi,ti f xi,ti 2 ,

Lbnd = 1 Nb j=1 Nb B uθ xj,tj g xj,tj 2 ,

Linit = 1 N0 k=1 N0 uθ xk,0 uθ xk 2 .

Here, Lres , Lbnd , and Linit denotes PDE residual loss, boundary condition loss, and initial condition loss, respectively, to ensure consistency, stabilization and smoothness. Nr , Nb , N0 denotes the total number of residual points, boundary points, and initial points, respectively. The weighting parameters λb,λ0 ,λreg >0 , are positive scaling coefficients that control the relative importance of boundary, initial, and regularization losses, respectively.

Since the basis functions φn ξ x,t are polynomials of the transformed coordinate ξ x,t , their derivatives with respect to t and x can be computed analytically through the chain rule. Hence,

(4.6)
t uθ x,t = n=0 Ncn θ φn ξ tξ,

(4.7)
x uθ x,t = n=0 Ncn θ φn ξ xξ,

If cn depends on x,t , the polynomial structure reduces numerical error in derivative evaluation relative to generic activation because φn ξ are known polynomials of degree n1 !. When the coefficient functions cn are allowed to depend on both spatial and temporal variables, i.e., cn=cn x,t;θc , the spectral representation becomes semi-local, combining data adaptability with the analytical structure of the orthogonal basis.

2.8. Network initialisation

An MNP-PINN of depth l consists of layers

(5.1)
h l =φnl z l ,

where

z l =W l h l1 +b l ,

with h 0 =x d0 ,W l dl×dl1, and b l dl , where we choose distributions for W l and b l such that

  • i.

    h l neither collapses to 0 nor blows up (non-degenerate forward pass).

  • ii.

    Gradients L/W l remain well-scaled across layers.

  • iii.

    The MNP orthogonality structure and its weighted normalisation are preserved (at least initially), so polynomial derivatives behave as expected.

Let’s analyse the propagation of the activation statistics through layers. Assume input xi have zero mean and variance σx2 , and weights are independent and identically distributed i.i.d. random variables with

(5.2)
E Wi,j l =0,Var Wi,j l =σw2 .

Then, for neutron i in layer l, we have that

(5.3)
zi l = j=1 dl1, Wi,j l hj l1 +bi l .

Assume hj l1 are i.i.d. with E hj l1 =0 and Var hj l1 =ql1 . Also let, Var bi l =σb2 . Then by independence, we get

(5.4)
E zi l =0, Var zi l =dl1, σw2 ql1 +σb2 .

For a nonlinear activation, let the activation at layer l be h l =φnl z l . Define the second moment map as

M 2,n σz2 =EZN 0,σz2 φn2 z .

Then the variance of post-activation is

(5.5)
ql=Var hi l =M 2,n dl1, σw2 ql1 +σb2 .

For a symmetric input distribution, we can approximate using even-order coefficients of φn , given as

(5.6)
Var h l k=1 nan,k 2 E z l 2k ,

where an,k are the MNP coefficients.

If z l N 0,σz2 , then

E z l 2k = 2k1 !!σz 2k .

To prevent variance blow-up, we impose energy conservation between layers by setting

(5.7)
Var h l =Var h l1 .

Substituting (5.7) into ((5.6) gives

dl1 σw2 k=1 nan,k 2 2k1 !!σz 2k2 =1.

Hence,

(5.8)
σw2 = 1 dl1 k=1 n an,k 2 2k1 !!σz 2k2 ,

which defines the critical weight variance to ensure stability. To this effect, we set σz2 =σx2 for the first layer. Deeper layers can be recursively estimated.

For instance, let us consider a special case involving the second-order MNP φn x= 1 3 5 x2 2 . We have the coefficients a 2,2 = 5 3 , a 0,2 = 2 3 . Then,

Var h l 5 3 2 E z4 = 25 3 σz4 .

Hence, the critical initialisation becomes

σw2 = 1 dl1 25 3 σz2 .

Setting σz2 =1 arbitrarily, we have

σw2 = 3 25 dl1 ,

which ensures stable propagation for quadratic MNP activations. Also, quadratic activations magnify magnitudes more strongly, so weights must be initialised smaller to maintain stable variance.

3. RESULTS AND DISCUSSION

3.1. Numerical example

We demonstrate the accuracy of the proposed method by considering a standard one-dimensional Poisson problem. Let u: 1,1 satisfy

u x=π2 sin πx ,x 1,1 ,

subject to the homogeneous Dirichlet boundary conditions

u 1 =u 1 =0.

The analytical solution of this boundary value problem is

ux=sin πx .

This example is smooth and well-posed, and therefore serves as a suitable benchmark for testing both accuracy and convergence behaviour of the proposed Mamadu–Njoseh-based Physics-Informed Neural Network (MNP-PINN) framework. The aim is to approximate ux numerically over 1,1 and compare the obtained solution with the exact one in terms of the L2 ​ and L error norms. The coefficients cn,n=0,,N, the lifted-basis multiplier lx , and the training parameters are chosen to effectively control the dynamics of the equation.

All network parameters are initialised according to Table 1 to ensure stable and efficient training:

Table 1: Initialisation parameters.
Parameter Value Description
Polynomial degree n=2 Quadratic Mamadu–Njoseh Polynomial (MNP) basis
Input dimension d=1 One-dimensional input
Target preserved variance q=1 Aim for unit activation variance
Pre-activation bound R1 Conservative bound to keep
Hidden neurons m=40 Number of neurons in the hidden layer
Interior collocation points Ninit =200 Interior training/collocation points
Boundary points Nbc =2, located at x=1,1. Boundary training points
Output-weight init. std. dev. σθ=1× 10 2 Standard deviation for output weight initialisation
Bias initialisation bj=0 All biases are initialised to zero
Weight initialisation Critical variance initialisation for quadratic MNP Use the derived optimal variance for stability
Scaling coefficient λreg = 10 6

Based on the initialisation and configuration parameters summarized in Table 1 for the MNP-PINN framework, the ensuing results are obtained via Python implementation and are presented in the subsequent tables and figures.

The network is trained using the Adam optimiser with an initial learning rate of 1× 10 6 , decaying exponentially by a factor of 0.5 every 1000 epochs. The total training duration is 3000 epochs. Loss function and error convergence at selected epochs are reported in Table 2, demonstrating rapid convergence and high-accuracy approximations with L2 ​ and L ​ errors on the order of 10 5 .

Table 2: Training snapshots of Mamadu-Njoseh polynomial physics informed neural network (MNP-PINN) for the 1D poison problem.
Epoch Loss (×10⁰) L₂ Error L∞ Error
0 4.854 × 101 9.440 × 101 9.926 × 101
500 2.520 × 10 1.600 × 10 2.050 × 10
1000 6.402 × 102 7.337 × 10 1.137 × 103
1500 2.657 × 10 1.565 × 10 1.988 × 10
2000 2.620 × 10 1.422 × 10 1.749 × 10
2500 7.875 × 10 2.343 × 10 3.578 × 10
2999 4.558 × 10 1.075 × 10 1.586 × 10

Collocation points are sampled uniformly within the computational domain 1,1 , resulting in 200 interior points. Boundary points are fixed at the endpoints x=1 and x=1 , enforcing Dirichlet boundary conditions. This procedure provides sufficient coverage of the domain for the 1D Poisson problem while maintaining computational efficiency.

3.2. Discussion

Figure 2 depicts the performance of the MNP-PINN applied to the 1D Poison Problem, illustrating the model’s ability to approximate the analytical solution of the target problem with remarkable precision. The figure typically shows both the predicted solution obtained by the MNP-PINN and the exact analytical solution, often plotted together for direct visual comparison. The near-perfect overlap between predicted solution and exact solution across the entire spatial domain demonstrates the spectral accuracy of the MNP-PINN. This high degree of correspondence implies that the MNP activation functions successfully capture the underlying smoothness and structure of the solution, even in regions with steep gradients or boundary effects. Additionally, the residual error often visualised below the main plot remains close to zero throughout, confirming numerical stability and robust convergence during training. The network’s ability to generalise well on such benchmark examples underscores the effectiveness of MNP layers in achieving physics-consistent approximations for differential problems, outperforming traditional neural architectures that rely on standard nonlinear activations.

Mamadu-Njoseh polynomial physics informed neural network (MNP-PINN) for 1D poison problem. X represents the independent variable, u(x) represents the dependent variable. MNPNN: Mamadu–Njoseh polynomial neural network..
Figure 2:
Mamadu-Njoseh polynomial physics informed neural network (MNP-PINN) for 1D poison problem. X represents the independent variable, u(x) represents the dependent variable. MNPNN: Mamadu–Njoseh polynomial neural network..

Table 2 summarizes the training performance of the MNP-PINN across selected epochs for the 1D Poisson equation. The table records the evolution of the loss function, the L2 norm error, and the L∞ norm error throughout the optimisation process. The results show a progressive and stable convergence pattern. Initially, the loss begins at a relatively high value (4.854×101) and decreases by nearly seven orders of magnitude, reaching values around 10 −6 after sufficient training. This sharp decline signifies efficient error minimization and well-conditioned gradient flow within the MNP-PINN framework. Both the L2​ and L∞​ error norms exhibit consistent reduction, converging to approximately 10−5, which corresponds to spectral-level accuracy, which is a hallmark of high-order polynomial approximation. A minor oscillation in the error curve around epoch 1000 reflects transient optimiser nonlinearity or adaptive learning rate adjustment, but the model rapidly stabilizes thereafter, reaffirming its robust training dynamics. Overall, the table confirms that the MNP-PINN achieves high numerical precision, strong generalisation, and stable convergence behaviour, validating the effectiveness of MNP activation structures in solving partial differential equations like the 1D Poisson problem.

Figure 3 presents the coefficient spectrum of the MNP-PINN, showing the magnitude decay of the polynomial coefficients cn as a function of the polynomial degree n, this spectrum provides insight into the representational efficiency and smoothness of the learned solution. As n increases, the coefficients cn exhibit a rapid exponential-like decay, indicating that the MNP-PINN effectively captures the dominant low-order modes of the solution while suppressing higher-order components. This behaviour confirms spectral convergence, meaning that the error decreases faster than any algebraic rate as the polynomial order increases. Such rapid coefficient attenuation signifies that higher-degree Mamadu–Njoseh basis functions contribute negligibly to the final approximation, thereby ensuring a compact spectral representation. This property reduces overfitting and numerical oscillations, common issues in classical neural architectures using purely analytical activations. Consequently, the MNP-PINN achieves smooth, stable, and physically consistent approximations across the spatial domain, showcasing the advantage of embedding Mamadu–Njoseh orthogonal structures within the neural framework.

Coefficient spectrum of Mamadu-Njoseh polynomial physics informed neural network (MNP-PINN) for the 1D poison problem.
Figure 3:
Coefficient spectrum of Mamadu-Njoseh polynomial physics informed neural network (MNP-PINN) for the 1D poison problem.

Figure 4 presents the pointwise absolute error distribution between the MNP-PINN-predicted solution u^(x) and the analytical solution u(x) across the spatial domain. The error is defined as

Ex= u^xux ,

Pointwise absolute error of Mamadu-Njoseh polynomial physics informed neural network (MNP-PINN) for the 1D poison problem. x represents the independent variable.
Figure 4:
Pointwise absolute error of Mamadu-Njoseh polynomial physics informed neural network (MNP-PINN) for the 1D poison problem. x represents the independent variable.

where u^x is the computed solution, and ut is the exact solution. The plot reveals that the MNP-PINN achieves uniformly low errors throughout the domain, with the maximum deviation confined to the boundary regions where solution gradients are typically steeper. The central region maintains near-zero error, highlighting the network’s strong approximation capacity and stability. The observed smoothness and near-symmetry of Ex reflect the polynomial orthogonality and spectral convergence characteristics of the Mamadu–Njoseh basis functions. Overall, the error magnitude remains on the order of 10−5, confirming high numerical accuracy and demonstrating the capability of MNP-PINN to resolve fine-scale spatial structures in the Poisson solution with minimal numerical dispersion.

Figure 5 illustrates the evolution of the MNP-PINN during training. The top panel shows the training loss decreasing exponentially by several orders of magnitude, confirming rapid and stable convergence. The lower four panels present MNP-PINN predictions at selected epochs (initial, early, mid, and final), revealing how the network progressively refines its approximation until achieving near-perfect overlap with the analytical solution. Minor oscillations in the loss reflect optimiser nonlinearity but do not affect overall stability. By the final epoch, both L2​ and L∞​ errors reach the order of 10−5, demonstrating spectral-level accuracy and excellent generalisation. The figure highlights the MNP-PINN’s ability to efficiently learn smooth PDE solutions through polynomial-based activation structures.

Training progression of Mamadu-Njoseh polynomial physics informed neural network (MNP-PINN) for the ID poison problem. u(x) represents the dependent function. x represents the independent function.
Figure 5:
Training progression of Mamadu-Njoseh polynomial physics informed neural network (MNP-PINN) for the ID poison problem. u(x) represents the dependent function. x represents the independent function.

4. CONCLUSION

The numerical experiments clearly establish the effectiveness of the MNP-PINN for solving the one-dimensional Poisson problem. The solution plots demonstrate an almost perfect overlap between the MNP-PINN predictions and the analytical solution, confirming spectral-level accuracy and strong physics consistency. Quantitative metrics further support this finding: the loss function decreases by several orders of magnitude, while both L2 and L errors converge to the order of ( 10 5 ), highlighting efficient optimisation and stable gradient dynamics.

The coefficient spectrum reveals rapid exponential decay in the polynomial coefficients, validating the compactness and representational efficiency of the MNP basis and indicating that only low-order modes are required to accurately approximate the solution. Pointwise error profiles remain uniformly small across the domain, with slight elevations near boundaries consistent with steep gradient regions, further demonstrating smooth and well-behaved convergence. Training-trajectory visualisations reinforce these findings, showing progressive refinement toward the exact solution with stable optimisation despite minor transient oscillations.

Hence, the results confirm that integrating MNP polynomial structures within the PINN framework yields highly accurate, stable, and spectrally convergent solutions. Thus, MNP-PINN represents a powerful and reliable methodology for solving differential equations, outperforming traditional PINN architectures that rely solely on standard nonlinear activations and demonstrating strong potential for broader PDE applications.

Ethical approval

Institutional Review Board approval is not required.

Declaration of patient consent

Patient's consent not required as there are no patients in this study.

Financial support and sponsorship

Nil

Conflicts of interest

There are no conflicts of interest.

Use of artificial intelligence (AI)-assisted technology for manuscript preparation

The authors confirm that there was no use of artificial intelligence (AI)-assisted technology for assisting in the writing or editing of the manuscript and no images were manipulated using AI.

REFERENCES

  1. , et al. Journal of Computational Physics. 2019;378:686-707.
  2. , et al. Nature Reviews Physics. 2021;3:422-40.
  3. , et al. Nature Computational Science. 2022;2:325-38.
  4. , et al. Journal of Computational Physics. 2020;425:109913.
  5. , et al. Water Resources Research. 2020;56:e2019WR026731.
  6. , et al. Journal of Computational Physics. 2021;436:110207.
  7. , et al. SIAM Review. 2021;63:208-2.
  8. , et al. SIAM Journal of Scientific Computing. 2022;44:A1526-A1553.
  9. , et al. Computers Mathematics with Applications. 2023;126:96-112.
  10. , et al. Computer Methods in Applied Mechanics and Engineering. 2022;389:114333.
  11. . Neural Networks. 2021;135:192-200.
    [CrossRef] [PubMed]
  12. , et al. Computers Mathematics with Applications. 2022;104:233-52.
  13. , et al. AIMS Mathematics. 2024;9:12775-777.
  14. , et al. Neural Networks. 2021;136:176-88.
  15. , et al. Science World Journal. 2016;11:21-24.
  16. , et al. Springer. 2004;44:165-179.
  17. . Journal of Mathematical Analysis and Applications. 1995;192:142-60.
  18. . Introduction to numerical solution of partial differential equations. In: Programming phase-field modeling. Springer; . p. :9-11.
    [Google Scholar]
  19. , et al. Mathematics of Computation. 1992;59:145-62.
  20. . The Finite Element Method for Elliptic Problems. Amsterdam: North Holland; .
  21. . A Mathematical Journal. 2004;6:1-32.
  22. , et al. Journal of Scientific Computing. 2024;100:35.
Show Sections