Process of Learning from Demonstration with Paraconsistent Artiﬁcial Neural Cells for Application in Linear Cartesian Robots

: Paraconsistent Annotated Logic (PAL) is a type of non-classical logic based on concepts that allow, under certain conditions, for one to accept contradictions without invalidating conclusions. The Paraconsistent Artiﬁcial Neural Cell of Learning ( l PANCell) algorithm was created from PAL-based equations. With its procedures for learning discrete patterns being represented by values contained in the closed interval between 0 and 1, the l PANCell algorithm presents responses similar to those of nonlinear dynamical systems. In this work, several tests were carried out to validate the operation of the l PANCell algorithm in a learning from demonstration (L f D) framework applied to a linear Cartesian robot (gantry robot), which was moving rectangular metallic workpieces. For the L f D process used in the teaching of trajectories in the x and y axes of the linear Cartesian robot, a Paraconsistent Artiﬁcial Neural Network ( l PANnet) was built, which was composed of eight l PANCells. The results showed that l PANnet has dynamic properties with a robustness to disturbances, both in the learning process by demonstration, as well as in the imitation process. Based on this work, paraconsistent artiﬁcial neural networks of a greater complexity, which are composed of l PANCells, can be formed. This study will provide a strong contribution to research regarding learning from demonstration frameworks being applied in robotics.


Introduction
Innovation in the field of automation requires research into the area of artificial intelligence, which includes other sub-areas to be studied in their individual or correlated forms. Among these, we can cite research that is aimed at making machines more independent from programming processes through expert systems, autonomous robotics, and machine learning [1][2][3]. In this context, this article shows new alternatives in automation control that require machine learning to add a higher level of efficiency by implementing algorithmic structures based on paraconsistent logic-PL [4][5][6]. π tutor : Ω → a ∈ A. (1) The Learner disposes of a set of observations O (named observation space) about the world and its own set of actions B.
The Learner follows another policy π learner in order to produce a behavior similar to the observed one, therefore π learner : O → b ∈ B, π learner ≡ π tutor . (2) Due to the condition that the Tutor and the system are involved in the same world, there is a perception equivalence problem in Lf D.
Since the interaction between the Tutor and the system takes place in the real world, information is obtained by each of the actors through different means.
The world is observed by the system through sensors, whereas the human observes it through its own senses. In this way, the Tutor demonstrating a particular behavior can observe modifications of the world that the Learner cannot perceive. Therefore, with the comparisons of information data between the Tutor and the system contaminated with contradictions, thus generating uncertainties in the conclusions, the Lf D process can reach a low level of efficiency [10,11].
Considering the conditions in which the Lf D processes operate, the application of PL, which has the property of supporting contradictory information, can alleviate some problems related to this technique. In this paper we present the application of the Paraconsistent Artificial Neural Cell (lPANCell) algorithm based on Paraconsistent Annotated Logic (PAL) integrated with the learning from demonstration process (Lf D) [12][13][14].

Related Works
The presentation of task models in such a way that a robot is able to learn from demonstration (Lf D)-that is, without the need for new programming-is a challenge for the field of robotics engineering. Several published articles bring results from research on learning from demonstration (Lf D), which aims to transfer task models to a robot through demonstration. For example, in [8] the authors present several Lf D approaches to allow a robot to learn and to generalize complex tasks from demonstrations.
In this article, we show a way that enables the learning of complex trajectories with complex sets of states, which the Cartesian robot needs to reach in a certain sequence to accomplish a task. Among the works related to the approach we developed in this research, we can highlight the studies presented in [15][16][17][18][19], where the authors present approaches in which robots learn task models through interaction with human tutors, and which could identify the aim simply as a difference between the initial and final states. In other approaches, as seen in [20], the task model includes not only the objective, but also a set of ordering constraints between actions.
Other important works in the Lf D process present classical artificial intelligence (AI) techniques for knowledge representation, planning, and learning. Several methods have been proposed in AI to improve planners by acquiring different types of knowledge of task model planning. Among these we can cite the methods published in [21]. In [22,23], the authors dealt with learning procedures that are related to planning speed and individual action attempts, which were often used together.
Some works, such as the one seen in [24], the authors addressed the method of inferring a generalized plan that works in all instances with a class of problems, efficiently instantiating plans for certain types of problems.
Other works, as seen in [25], focused on learning Hierarchical Task Networks (HTN) that are used to hierarchically represent the planning knowledge about a problem domain. As demonstrated in several articles, for example in [26], planning task models and transferring them efficiently faces two main problems. The first is that, when the task is complex, some extra knowledge about how to plan and create adaptive models capable of being reproduced by the robot is needed. Additionally, the second is that, when there are different alternatives to reach a goal, modeling can be performed under conditions that are based on different constraints, mainly physical ones. Therefore, task planning systems often require generic task models, ones which specify how tasks are to be correctly achieved.
In this work, we show a method based on artificial intelligence techniques that are supported by non-classical logic; a method that brings algorithms and configurations to be applied in model planning and that considers a robot equipped with a set of basic skills. Thus, the contribution of this article will be a technique that can be applied based on algorithms that are based on Paraconsistent Annotated Logic [6,27].
The usual symbols and PAL nomenclature used in this article are shown in Table 1. As will be described in this article, we can build different configurations with structures composed of Paraconsistent Artificial Neural Networks [28]. Thus, we developed Lf D strategies for representations dedicated to the modeling of tasks, ones that are represented in complex trajectories.

Organization
This article presents the following organization: In addition to this introduction, we present, in Section 2, the main concepts of the method of learning from demonstration, highlighting the techniques that will be used in the tests and applications of Paraconsistent Neural Cells. In Section 3, we present the fundamentals of Paraconsistent Annotated Logic (PAL), as well as its equations and algorithms that form the learning Paraconsistent Artificial Neural Cell (lPANCell). We finish Section 3 with the presentation of a network algorithmic structure (lPANnet) composed of lPANCells that are interconnected in a cascade. In Section 4, which refers to materials and methods, we present the techniques and procedures used to build the lPANCell configurations and its implementation, thus forming a lPANC_BLK block that follows the rules of IEC 61131-3; furthermore, this section is dedicated to the learning from demonstration process. We conclude, in Section 4, by presenting, with a network structure of lPANC_BLK blocks (lPANnet), the procedures that are used by the linear Cartesian robot regarding the automatic machines that are used in the tests of moving rectangular metal workpieces. In Section 5, we present the graphs and tables that show the results obtained in this work related to the performance of the Paraconsistent Artificial Neural Network (lPANnet) structure that is used in the process of moving workpieces (which is carried out by the machine tool). In Section 5, we also discuss the simulation results and the expected and obtained values in the workpiece movement process. In Section 6, we present the conclusions and the possibilities for future work applying this technique.

Fundamentals of Learning from Demonstration (Lf D)
In systems dedicated to the control of automation, trajectories, and robot movement, highly complex algorithms are used. The programming of the computational modules of these machines uses dedicated and structured algorithms so that the machines can perform tasks that are related to the movement and transport of objects with precision [29]. However, with each change in the machine trajectory or changes in transported products, there is a need for the intervention of programming specialists to adapt the system to these new tasks [29][30][31]. In industrial production systems where robots act in complex actions, reprogramming to adapt the machines can make production unfeasible. Thus, learning from demonstration (Lf D) would be a method that would provide greater speed in modifying machine movement strategies, where, in this case, the knowledge transfer process would use the concepts of human learning. Thus, flexibility in machine programming would have to be adapted with algorithms capable of transferring skills through direct intervention, observation, goal emulation, imitation, and other social interactions [32][33][34][35].

Machine Learning
Machine learning is used to teach machines how to handle the data more efficiently. As can be seen in [36], machine learning methods can be classified into four types: 1. supervised learning; 2. unsupervised learning; 3. semi-supervised learning; and 4. reinforced learning.

Supervised Learning
Supervised learning, which is considered to be Lf D with the support of a system composed of artificial neural networks, can be classified into three stages: demonstration, learning, and imitation. These steps will be described below [36,37].
(1) Demonstration-This is where the Tutor performs the desired activities for the robot to learn. In a practical process, it is considered that the Learner and the Tutor are exposed to a training vector. The Tutor, based on his knowledge of the environment, has the ability to provide a response to the artificial neural network, which represents the optimal action to be performed by the learner according to the training vector; (2) Learning-In this step supervised learning takes place. This is where the parameters of the artificial neural network are adjusted based on the training vector and on the error signal, which is defined by the difference between the desired response and the real response of the network. In this step, the artificial neural network simulates the Tutor, transferring the knowledge of the environment to the Learner. In general, the optimization of the emulation is verified through statistical calculation to confirm that the knowledge of the environment is transferred to the learner in a complete way; (3) Imitation-At this stage, the learning condition is complete, and the artificial neural network has finished its training process transferring the knowledge obtained from the Tutor to the Learner. Therefore, the Tutor is no longer needed, and the Learner, after receiving the information from the artificial neural network, is able to recognize and deal with the environment without the need for external help. Figure 1 shows a block diagram on the main steps of supervised learning [37,38].
exposed to a training vector. The Tutor, based on his knowledge of the e the ability to provide a response to the artificial neural network, whic optimal action to be performed by the learner according to the training v (2) Learning-In this step supervised learning takes place. Th parameters of the artificial neural network are adjusted based on the tra on the error signal, which is defined by the difference between the desir the real response of the network. In this step, the artificial neural netwo Tutor, transferring the knowledge of the environment to the Learner optimization of the emulation is verified through statistical calculation to knowledge of the environment is transferred to the learner in a complete (3) Imitation-At this stage, the learning condition is complete, a neural network has finished its training process transferring the knowledg the Tutor to the Learner. Therefore, the Tutor is no longer needed, and t receiving the information from the artificial neural network, is able to rec with the environment without the need for external help. Figure 1 shows a block diagram on the main steps of supervised lea The "Learning Techniques" block in Figure 1 is related to m techniques, therefore it is where we will use an LfD method that Paraconsistent Logic algorithms. The "Learning Techniques" block in Figure 1 is related to machine learning techniques, therefore it is where we will use an Lf D method that is applied with Paraconsistent Logic algorithms.

Demonstration Step Approaches
In general, three forms of approaches are used for the demonstration stage. The first is known as kinesthetic, which is where the Tutor physically demonstrates the moving actions to the robot or machine through the desired movements or trajectories. In this case, the robot states during the interaction are registered through its integrated sensors (e.g., joint angles and torques), resulting in training information data for the machine learning model [37,38].
The second form of demonstration is called teleoperation. In this category, there is no requirement for the Tutor to be in direct contact with the Learner (which, in this case, is the robot). For this demonstration technique, an external input is required to drive the machine or robot; this can be via a joystick, a graphical interface, or other means. This mode allows Lf D techniques to be applied in remote environments, and information about the robot's states during movement is registered through its integrated sensors.
The third demonstration approach is the Passive Observations or Motion-Sensor Demonstration method. In this approach, the Tutor performs the task using their own body, which can sometimes be implemented with additional sensors to facilitate tracking [36][37][38]. Information data about the machine's states and trajectories are registered by means of tracking sensors.

Data Modeling for Feature Extraction
The demonstration learning technique brings numerous challenges to its application; among these main problems is the data modeling mode for the extraction of features. This phase encompasses two main issues that need to be discussed. The first is how to interpret and understand the data arising from observed behaviors-in other words, the problem is how to recognize human behavior (Tutor) from information data. To achieve the robot (Learner) learning from the demonstration, it is necessary to address the matching problem, which means finding ways to map links and joints from a human (Tutor) to a robot (Learner) [39,40]. The second is how to integrate the systems of perception and movement control to reconstruct what is observed. Therefore, methods must be found to structure the motor control system for general movement and imitation learning capabilities. Faced with these challenges, current approaches are divided into two trends: the first uses trajectory coding, which is a low-level representation of skill, in the form of a non-linear mapping between sensory and motor information; the second trend has used symbolic coding, which is a high-level representation of a skill that breaks down a skill into a sequence of action perception units [41,42].

Modeling
In this learning from demonstration system, generalization is important since it should be possible to model a demonstrated movement with different goal positions. Another important feature is that a learning from demonstration model must also be robust against disturbances, as exactly repeating an observed movement is not realistic in a dynamic environment in which obstacles can suddenly appear. Due to this and other factors, the data collected in the demonstration approaches need to be mapped through a mathematical model to compose the algorithmic models that will enable imitation acts. Therefore, systems operating in Lf D require a learning policy to extract and generalize the main features of assembly movements [42]. In [39,40], Hidden Markov Models (HMMs) were used to code and generalize the demonstration examples. In the normal Markov model, the state is directly visible to the observer; therefore, the transformation probabilities between states are the whole parameters. While the states are not visible in the Hidden Markov Model, some variables that are influenced by the states are visible. Every state has a probability distribution on a token of possible output; therefore, the sequence of the output token reveals the information of the state's sequence. The Hidden Markov Model (HMM) is a robust probabilistic method that is used to encapsulate human movement; it contains both spatial and temporal variables that are used for numerous demonstrations [40].
Another modeling method that has shown good results was included in [41][42][43], where the authors used dynamic motion primitives (DMPs), which represent a fundamentally different approach to representing motion, one that based on nonlinear dynamical systems.
Dynamical systems with global asymptotic stability define a function that represents a global navigation map where all space trajectories converge to the target [41]. Therefore, when used to code trajectories, this characteristic of dynamic systems ensures that every trajectory, within the regions of asymptotic stability, collaborates with the correct direction to reach the target from the robot's current position. If the robot or machine suffers any spatial disturbance that is caused by an actuation or sensing error that introduces small deviations in the execution of its trajectory, it will still be able to reach the target if the deviations do not remove it from the stability region [43]. Modeling with this technique also allows the system to be able to deal with the temporal disturbances that are caused by communication delays and the dynamics of low-level controllers. An alternative for the system to be robust to these temporal disturbances is that it be autonomous, i.e., independent of time [39,42,43].
As seen in [41][42][43], DMP is robust to spatial perturbations and is suitable for the purpose of following a specific motion path, so it is one of the modeling techniques that has responded robustly to perturbations in Lf D architectures.
In this article, we aim to present an algorithm based on Paraconsistent Logic, the Learning Paraconsistent Artificial Neural Cell-lPANCell [44] algorithm, which holds the characteristics of modeling by DMP, which is applied in the process of learning the range movements of robotic machines from a set of demonstrations [41,45]. Each demonstration is a trajectory in the linear Cartesian robotic machine's workspace, which has the task of moving rectangular metal workpieces in an industrial production process [46].
The information data corresponding to the trajectories are obtained through the robotic machine being controlled by a joystick operated by a human, where the positions are collected and transformed into learning patterns. After learning, the robotic machine will be able to reproduce, without human interference, trajectories that are similar to the demonstrations. In the imitation stage, the robotic machine, which has learned through the Lf D process, should show robustness to the temporal disturbances caused by communication delays, as well as to the spatial disturbances caused by some actuation or sensing error. The trajectories are similar to the demonstrations that are executed by the robot, even when starting the movement from points that are different from the demonstrations, and even with the change in the target position occurring during the execution of the procedure [46,47].

Paraconsistent Logic (PL)
Paraconsistent Logic (PL) belongs to a non-classical logic category, and its main feature is the revocation of the principle of non-contradiction [48]. In [49], PL was described with its equations and all its predicates being formalized; it was studied in completeness, and was demonstrated to be an example of non-classical, propositional, and evidential-based logic.
In [6], the authors presented an extension of PL called Paraconsistent Annotated Logic (PAL), which is associated with a four lattice system, wherein its vertices are considered representations of the following logical states: t = true, f = false, T = inconsistent, and ⊥ = paracomplete. In this representation, the logical states refer to a proposition P that is under analysis. In this mode of interpretation, an annotation composed of two degrees of evidence (µ, λ) assigns a logical connotation to a proposition P, such that µ represents the degree of evidence in favor of proposition P, and λ represents the degree of evidence against proposition P.
The degrees of evidence are extracted from different information sources and have their values normalized, thus belonging to a set of real numbers. Therefore, µ, λ ⊂ [0, 1] ∈ R.
Thus, the annotation (µ, λ), in the condition of maximum values, represents the extreme logical states of the lattice vertices, which are according to the values of µ and λ below: (1,0)-favorable evidence degree to proposition P maximum and unfavorable evidence degree to proposition P minimum resulting in a logical state t = true; (0,1)-favorable evidence degree to proposition P minimum and unfavorable evidence degree to proposition P maximum resulting in a logical state f = false; (1,1)-favorable evidence degree to proposition to P maximum and unfavorable evidence degree to proposition P maximum resulting in a logical state T = inconsistent; (0,0)-favorable evidence degree to proposition P minimum and unfavorable evidence degree to proposition P minimum resulting in a logical state ⊥= paracomplete.
As the values of µ and λ can vary between 0 and 1 and belong to the set R, then infinite logical states, with their variations, can be considered internally in the lattice.
Algebraic interpretations can relate to the degrees of evidence used in the logical states of the PAL2v lattice.
Through the representations of values, with favorable evidence degree µ on the x axis and unfavorable evidence degree of λ on the y axis in a Unit Square in the Cartesian Plane (USCP), it is possible to obtain a transformation (T), shown in Equation (3), that represents these values in a lattice that is associated with PAL [6,49,50].
Relating the lattice associated with PAL with components of the transformation T(X, Y) from Equation (3), we have The first term obtained in the ordered pair of the Equation (3) which is called the degree of certainty (Dc), and the second term is Y = x + y − 1 = µ + λ − 1 →, which is called the degree of contradiction (Dct). Therefore, the degree of certainty is obtained by Equation (4).
and the degree of contradiction is obtained by Equation (5).
The resulting evidence degree (µ E ) is calculated by the degree of certainty normalization, as is the case in Equation (6): Additionally, the normalized contradiction degree (µ ctr ) is calculated by µ ctr = Dct+1 2 or by Equation (7): With these PAL2v-equations, we can build algorithms for the analysis and comparison of signals through paraconsistent logic [4,6,49].
Based on Equations (5) and (6), a pair of values (Dc, Dct) is formed that provides a single point located within the lattice associated with PAL, thus defining a paraconsistent logical state [50][51][52]. Figure 2 shows a Unit Square in the Cartesian plane (USCP) with the evidence degrees µ and λ exposed on the x e, and y axes, and with the PAL lattice with a paraconsistent logic state composed by the pair (Dc, Dct).
The degrees of evidence originate from different sources, vary in the range between 0 and 1, and belong to a set of real numbers. Thus, this annotation of two values (µ, λ) can generate infinite points and can consequently generate infinite logical states (Dc, Dct) in the associated PAL lattice [6,50,52].

Paraconsistent Artificial Neural Cell of Learning (lPANCell)
As the Paraconsistent Logic Annotated with Annotation of Two Values (PAL2v) model has the ability to handle information that is considered incomplete and contradictory, then some algorithmic structures built with its equations may indicate similarities with the biological mental process.
In [6], a family of PANCells-Paraconsistent Artificial Neural Cells-was presented to compose computational structures that are capable of showing results similar to some behaviors of the human brain. Through an algorithm configured with Equation (4), a basic Paraconsistent Artificial Cell (PACellb) was considered, which-from the signals received at the input and from external adjustment factors-makes the decision, based on levels of resulting values in your output. Using PACellb as a basis, Paraconsistent Artificial Neural Cells were built, which were of different types and with specific purposes [4,51,53]. Based on Equations (5) and (6), a pair of values (Dc, Dct) is formed that provides a single point located within the lattice associated with PAL, thus defining a paraconsistent logical state [50][51][52]. Figure 2 shows a Unit Square in the Cartesian plane (USCP) with the evidence degrees μ and λ exposed on the x e, and y axes, and with the PAL lattice with a paraconsistent logic state composed by the pair (Dc, Dct). The degrees of evidence originate from different sources, vary in the range between 0 and 1, and belong to a set of real numbers. Thus, this annotation of two values (μ, λ) can generate infinite points and can consequently generate infinite logical states (Dc, Dct) in the associated PAL lattice [6,50,52].

Paraconsistent Artificial Neural Cell of Learning (lPANCell)
As the Paraconsistent Logic Annotated with Annotation of Two Values (PAL2v) model has the ability to handle information that is considered incomplete and contradictory, then some algorithmic structures built with its equations may indicate similarities with the biological mental process.
In [6], a family of PANCells-Paraconsistent Artificial Neural Cells-was presented to compose computational structures that are capable of showing results similar to some behaviors of the human brain. Through an algorithm configured with Equation (4), a basic Paraconsistent Artificial Cell (PACellb) was considered, which-from the signals received at the input and from external adjustment factors-makes the decision, based on levels of values, result from the exit. Using PACellb as a basis, Paraconsistent Artificial Neural Cells were built, which were of different types and with specific purposes [4,51,53].
In this work, we use the Paraconsistent Artificial Neural Cell of Learning (lPANCell) model, which can be trained to learn patterns that are between the real values of 0 and 1. Initially, it was considered that the pattern in a lPANCell is defined as a binary digit, in which 1 is equivalent to the "true" logic state and the value 0 means a "false" logic state. From Equation (4), it is possible to obtain a single recurrence equation so that the degree of unfavorable evidence λ is obtained from the complement of the degree resulting from the output [6,[49][50][51][52][53]. Considering that 1 μ is the value of the learning pattern in the current state, and ( ) 1 Ek μ + is the value of the later state-that is, the learned value-then the In this work, we use the Paraconsistent Artificial Neural Cell of Learning (lPANCell) model, which can be trained to learn patterns that are between the real values of 0 and 1. Initially, it was considered that the pattern in a lPANCell is defined as a binary digit, in which 1 is equivalent to the "true" logic state and the value 0 means a "false" logic state. From Equation (4), it is possible to obtain a single recurrence equation so that the degree of unfavorable evidence λ is obtained from the complement of the degree resulting from the output [6,[49][50][51][52][53]. Considering that µ 1 is the value of the learning pattern in the current state, and µ E(k+1) is the value of the later state-that is, the learned value-then the lPANCell equation to obtain the Resultant Evidence Degree µ E is represented by Equation (8): and k as the number of iterations, and l F = Learning Factor. The lPANCell equation to obtain the normalized contradiction degree µ ctr is represented by Equation (9) [6,12]: The lPANCell is considered completely trained when µ E(k+1) = 1 and µ ctr(k+1) = 0. If the pattern applied to the cell input is unitary (µ 1 = 1), then the output will converge to µ E(k+1) = 1 with a monotonic variation, whose number of iterations (n) for the output to reach this maximum value will depend on the l F value [4,7,13].
Present the current state of learning (Equation (10)).
If µ E(k+1) = 1 → Do µ E(k+1) = µ 2 and return to step 3 9. Stop Figure 3 shows the symbol of the lPANCell, as well as the graphs with the curves corresponding to the outputs (µ (k+1) ) that were obtained with 12 iterations of the pattern µ 1 = 1, with learning factor values adjusted to l F = 1, l F = 0.8, l F = 0.6, l F = 0.4, and l F = 0.2.
Robotics 2023, 12, x FOR PEER REVIEW 11 of 29 Figure 3 shows the symbol of the lPANCell, as well as the graphs with the curves corresponding to the outputs (μ(k+1)) that were obtained with 12 iterations of the pattern μ1 = 1, with learning factor values adjusted to lF = 1, lF = 0.8, lF = 0.6, lF = 0.4, and lF = 0.2.
It was verified that after repeatedly receiving its input, the value of the pattern μ1 resulted in an ascending monotonic curve appearing, which demonstrates the cell's learning after the convergence in output μ(k+1) = 1.0. When the value of the learning factor is set to lF = 1, the lPANCell will need more steps (iterations) to complete a training; in this case, it is characterized as a natural learning capacity [6].
Convergence for patterns of values between 0 and 1 only happens for the learning factor set in lF = 1. With the learning factor set to lF = 1, and by applying its input μ1 to the It was verified that after repeatedly receiving its input, the value of the pattern µ 1 resulted in an ascending monotonic curve appearing, which demonstrates the cell's learning after the convergence in output µ (k+1) = 1.0.
When the value of the learning factor is set to l F = 1, the lPANCell will need more steps (iterations) to complete a training; in this case, it is characterized as a natural learning capacity [6].
Convergence for patterns of values between 0 and 1 only happens for the learning factor set in l F = 1. With the learning factor set to l F = 1, and by applying its input µ 1 to the value of 0 repeatedly, a descending monotonic curve appears that demonstrates the unlearning of the cell when the convergence reaches in output µ (k+1) = 0.0 [6]. Table 2 shows the values obtained in the simulation of an lPANCell with the learning factor set at l F = 1 in the sequential application of five patterns with different values. The test was conducted with 10 iterations each. Figure 4 shows the values obtained in the form of graphs [4,7,13].  When the lF value is other than 1, the output will not converge to the input pattern value if the u1 pattern is different from 1. In this case, adaptations must be made to the input values with a compensation in the output values [13]. When the l F value is other than 1, the output will not converge to the input pattern value if the u1 pattern is different from 1. In this case, adaptations must be made to the input values with a compensation in the output values [13].

lPANnet-Paraconsistent Artificial Neural Network
The lPANCell model can be interconnected and can form different configurations of a paraconsistent artificial neural network for the purposes of an analysis of information signals that are formatted into degrees of evidence.
As seen in [13], it is possible to build a configuration of lPANCells that are interconnected in a cascade, and where the output of the first lPANCell (µ (kn1) ) is the same pattern applied to the second, etc. In this case, the pattern applied to the first lPANCell will have a response that is twice as fast as the second lPANCell, etc. Figure 5a shows a structure of an lPANnet composed of four lPANCells that were configured in a cascade. Simulations were carried out where, initially, applications of Pattern µ1 = 0 were made in the first lPANCell; these consisted of an unlearning process and, soon after, pattern µ1 = 1 was applied for 12 iterations.    Figure 5b shows the results, in graph form, that were obtained in the learning process of the lPANnet configuration in 12 iterations. The values shown in Table 3 refer to the results obtained from the four outputs of the lPANCell that make up the configuration [6,7,13]. Table 3. Values of the results obtained in the training process of the configuration with four lPANCells interconnected in a cascade, as shown in Figure 5a.

Materials and Methods
In this work, the lPANCell algorithm was implemented, with non-classical logic characteristics, in a Programmable Controller (PC) in the form of a functional block (lPANC_BLK) to be used in the control device of a linear Cartesian robot (gantry robot). The programming of the function block followed IEC 61131-3 standards [54], which aim to standardize programming languages and its structure in programmable controllers.

Functional Block (lPANC_BLK)
In the construction of the function block (lPANC_BLK) that was based on the learning Paraconsistent Artificial Neural Cell algorithm (lPANCell), the Programmable Controller (PC) from the manufacturer SIEMENS model S7-1500 was used. The program was developed using SIEMENS (TIA Portal)-Totally Integrated Automation Portal software [55].
The internal programming of the function block (lPANC_BLK) was carried out based on the lPANCell algorithm, in which the Ladder language was chosen due to its wide use by professionals in the industrial automation area. Figure 6 shows the lPANC_BLK symbol that was built based on the IEC 61131-3 standard [54] and the corresponding nomenclatures between the function block and the lPANCell symbol.
To the left of the lPANC_BLK block are the inputs: EN (enable), input pattern u1, learning factor l F , reset signal, and refresh signal. The output variables, which are located to the right of the block, are ENO (enable output), degree of the evidence of the output (uE), and signaling signal (uE_Signaled). To provide a better analysis of the behavior of the lPANC_BLK block, the values of lF, u1, and uE were normalized in the range between 0 and 100. Therefore, these three values were proportional to the normalized values of lPANCell, which is between 0 and 1.
The function of the input and output variables of lPANC_BLK are described below: (a) EN: Enable the block when its logic level is 1; (b) u 1 : Input pattern to be learned by lPANC_BLK in the range 0 to 100; (c) lF: Learning factor that corresponds to an adjustment of how quickly the cell will learn the input pattern-adjusted in the range of 0 to 100; (d) Reset: Restart the block, assigning the value 50 to the output µ E . by professionals in the industrial automation area. Figure 6 shows the lPANC_BLK symbol that was built based on the standard [54] and the corresponding nomenclatures between the function bl lPANCell symbol. To the left of the lPANC_BLK block are the inputs: EN (enable), input learning factor lF, reset signal, and refresh signal. The output variables, which to the right of the block, are ENO (enable output), degree of the evidence o (uE), and signaling signal (uE_Signaled). To provide a better analysis of the the lPANC_BLK block, the values of lF, u1, and uE were normalized in the ran 0 and 100. Therefore, these three values were proportional to the normalize lPANCell, which is between 0 and 1.
The function of the input and output variables of lPANC_BLK are descr (a) EN: Enable the block when its logic level is 1; (b) u1: Input pattern to be learned by lPANC_BLK in the range 0 to 100; (c) lF: Learning factor that corresponds to an adjustment of how quickly learn the input pattern-adjusted in the range of 0 to 100; The lPANC_BLK block, when receiving a pulse on the reset input, generates at its output uE the value 50, which corresponds to the value 0.5 in the lPANCell algorithm. The cell is then ready to start learning (or unlearning).
(e) Refresh: Updates the input pattern after receiving a pulse at logic level 1; (f) ENO: Provides logic level 1 when block is active; (g) uE: Evidence degree, which corresponds to the block output in the range of 0 to 100; (h) uE_Signaled: Signaling with a value of 100 when the block ends learning (or unlearning).

Linear Cartesian Robot and Pneumatic Machine Tool
The lPANC_BLK block was installed in the Programmable Controller (PC) of a linear Cartesian robot (gantry robot), which operates in a pneumatic machine tool whose task is to move metal workpieces in an industrial system [46,47,55].
The structure of the linear Cartesian Robot (gantry robot) composing the pneumatic machine tool is shown in Figure 7.
We can see in Figure 7 that the pneumatic machine tool consists of a double-acting rodless cylinder structure (1), whereby the gantry mounts a sliding crossbar to trigger the movements that are performed on the X axis. In addition, a double-acting rodless cylinder is fixed to the previous cylinder (2), which has the function of driving the support assembly that moves on the Y axis, which is where a gripping device is coupled with magnetic workpieces that are driven by a hollow through-rod cylinder with a suction cup (3). This suction cup's function is to move two linear transducers X and Y (4) along the Z axis, as well as to operate an electropneumatic valve block (5), an air-compressed unit (6), and a set of terminals (7) for the electrical control interface.
The lPANC_BLK block was installed in the Programmable Controller (PC) of a linear Cartesian robot (gantry robot), which operates in a pneumatic machine tool whose task is to move metal workpieces in an industrial system [46,47,55].
The structure of the linear Cartesian Robot (gantry robot) composing the pneumatic machine tool is shown in Figure 7. We can see in Figure 7 that the pneumatic machine tool consists of a double-acting rodless cylinder structure (1), whereby the gantry mounts a sliding crossbar to trigger the movements that are performed on the X axis. In addition, a double-acting rodless cylinder is fixed to the previous cylinder (2), which has the function of driving the support assembly that moves on the Y axis, which is where a gripping device is coupled with magnetic workpieces that are driven by a hollow through-rod cylinder with a suction cup (3). This suction cup's function is to move two linear transducers X and Y (4) along the Z axis, as well as to operate an electropneumatic valve block (5), an air-compressed unit (6), and a set of terminals (7) for the electrical control interface.

Practical Application-Forms of Activation and Control
In practice, the lPANnet computational structure (composed of lPANC_BLK functional blocks) acts in an LfD process to teach trajectories (on the x and y axes) to a linear Cartesian robot (gantry robot) to be able to work on moving rectangular metallic workpieces on the table of an electropneumatic machine tool.

Practical Application-Forms of Activation and Control
In practice, the lPANnet computational structure (composed of lPANC_BLK functional blocks) acts in an Lf D process to teach trajectories (on the x and y axes) to a linear Cartesian robot (gantry robot) to be able to work on moving rectangular metallic workpieces on the table of an electropneumatic machine tool.
As shown in Figure 8, the machine contains a triangular table with dimensions 670 × 890 mm. This is where the metallic workpieces are moved through the action of the linear Cartesian robot (gantry robot). The dimensions that are useful for delimiting the trajectories x and y through which the linear Cartesian robot (gantry robot) works are 500 × 700 mm. Therefore, the drive assembly of the linear Cartesian robot, which is composed of the cylinder-driven sliding crossbar, can move in a 700 mm extension, which is the table length, on the X-axis. In addition, the support carrying the magnetic gripper is driven by another cylinder and can move 500 mm on the Y-axis. The magnetic workpiece gripper device captures one metallic workpiece at a time and is capable of moving around 20 mm on the Z axis.
For the Lf D process, the pneumatic machine tool has a manual push button (joystick), a programmable controller (PC), two displacement transducers, a flag device (5), and other support devices, as shown in Figure 8. The electronic system composed of the PC and electronic devices controls the movement of the workpiece magnetic gripping device from a determined point as the origin (X 0 , Y 0 ) to the place of deposition of the workpieces, which is stipulated as the objective point of the trajectory (X (Target) , Y (Target) ). Two transducers measure the distances traveled in moving the slider bar (X-axis transducer) and moving the workpiece magnetic gripping device support (Y-axis transducer).
The pushbutton (joystick) has movements on the XYZ axes, and its activation can occur manually through a remote control. The program inserted in the Programmable Controller (PC) was made in such a way that when putting the system into manual operation the lPANC_BLK Block will enter the learning mode. and electronic devices controls the movement of the workpiece magnetic gripping device from a determined point as the origin (X0, Y0) to the place of deposition of the workpieces, which is stipulated as the objective point of the trajectory (X(Target), Y(Target)). Two transducers measure the distances traveled in moving the slider bar (X-axis transducer) and moving the workpiece magnetic gripping device support (Y-axis transducer). The pushbutton (joystick) has movements on the XYZ axes, and its activation can occur manually through a remote control. The program inserted in the Programmable Controller (PC) was made in such a way that when putting the system into manual operation the lPANC_BLK Block will enter the learning mode.

Paraconsistent Artificial Neural Network (lPANnet) Configured with lPANC_BLK Blocks
In the Programmable Controller, a program was installed that forms a structure composed of a set of 8 lPANC_BLK blocks forming a Paraconsistent Artificial Neural Network (lPANnet) that will operate in the LfD process.

Paraconsistent Artificial Neural Network (lPANnet) Configured with lPANC_BLK Blocks
In the Programmable Controller, a program was installed that forms a structure composed of a set of 8 lPANC_BLK blocks forming a Paraconsistent Artificial Neural Network (lPANnet) that will operate in the Lf D process. Figure 9 shows the configuration of the lPANCells that were formed through the lPANC_BLK blocks.
Each group of 4 lPANC_BLK blocks interconnected in a cascade will be trained separately. The first group (lPANnet Y) receives the pattern from the Y transducer and learns the trajectory demonstrated by teleoperation on the y-axis.
The second group (lPANnet X) receives the pattern from the X transducer and learns the trajectory demonstrated by teleoperation on the x-axis.
The imitation step can be conducted by choosing one of the 16 possible trajectories in the x and y axes, established by the transition states (Xn, Yn) that are represented by the output values of the second group (lPANnet X) and first group (lPANnet Y), respectively.
The l F values of each lPANCell and the composition of the computational structure that make up the lPANnet allow for obtaining trajectory adjustments, as well as obtaining the velocities linked to the transition period of the states, thus resulting in linear or non-linear trajectories. output values of the second group (lPANnet X) and first group (lPANnet Y), respectively.
The lF values of each lPANCell and the composition of the computational structure that make up the lPANnet allow for obtaining trajectory adjustments, as well as obtaining the velocities linked to the transition period of the states, thus resulting in linear or nonlinear trajectories.

Planning
In this work, due to the characteristics of the machine tool, the trajectories were planned as shown below.

Planning
In this work, due to the characteristics of the machine tool, the trajectories were planned as shown below.
X-axis trajectory: start X 0 = 0 mm finish X (Target) = 700 mm Y-axis trajectory: start Y 0 = 0 mm finish Y (Target) = 500 mm If the learning factor is set to l F = 1, then 12 iterations are required for a complete learning of µ (k+1) = 1.
The advance corresponds to the resulting degrees of evidence and, in this way, the trajectory of the movement is fulfilled through the variations in the transition states. The l F adjustment values will influence the trajectory behavior where advances will be greater at the beginning and will then decrease at the end of learning.
The training process was carried out with the button selected in manual mode (teleoperation).
Although the learning was estimated to be completed after 10 iterations, the occurrence of interferences in learning can lead to more iterations being necessary for the output of lPANC_BLK to better indicate the completed learning value. Thus, the system was adjusted so that, after the necessary iterations, the lPANC_BLK indicated the maximum value of µ (k+1) = µ 1 with the "Trained Cell" signal. After this signal, the linear Cartesian robot, which drives the machine tool, was available for the "Imitation" step, where the system works without operator intervention.

The "Demonstration
Step" was performed in two stages:

1.
Learning from demonstration on the x-axis determining the trajectory gx = X (Target) − X 0 ; 2.
Learning from demonstration on the y-axis determining the trajectory gy = Y (Target) − Y 0 . This makes the linear Cartesian robot activate the support assembly where the magnetic workpiece gripper device is attached, which fulfills the moving of the workpieces in the trajectory that was established by the learning of advances in both axes.

Learning Stage (Teleoperation)
(a) First gx trajectory-Advance of the pneumatic cylinder on the x axis In possession of the button (joystick), the Tutor activates the machine tool (Learner) n times, causing the support with the magnetic grabber to capture the metallic workpiece at point X 0 and to move it to the indicated point X (Target) . With the movement of the magnetic grabber, the displacement transducer on the x axis brings the information of the cylinder advance in the form of an evidence degree (a value between 0 and 1), which is applied in the cells that are implemented in the lPANC_BLK x blocks.
In this step, the first lPANCell cell programmed in the lPANC_BLK x1 block learns the pattern corresponding to the advance distance of the trajectory gx, so its output after each iteration will present a value corresponding to the advance distance gx = X (Target) − X 0 . In complete learning, the value at the outputs of all 4 lPANCells will be maximum µ (k+n)xn = 1 that corresponds to the maximum permissible advance on the machine tool table (700 mm).
(b) Second gy trajectory-Advance of the pneumatic cylinder on the y axis The Tutor activates the machine tool (Learner) n times causing the magnetic gripper to capture the metallic workpiece at point Y 0 and move it to the indicated point Y (Target) . With the movement of the magnetic grabber, the displacement transducer on the y axis brings the information of the cylinder advance in the form of evidence degree (value between 0 and 1) that is applied in the cells implemented in the lPANC_BLK y blocks. In complete learning, the value at the outputs of all 4 lPANCells will be a maximum µ (k+n)yn = 1, which corresponds to the maximum permissible advance on the machine tool table (500 mm).
Completing the procedures to lPANnet, which comprises the 4 lPANC_BLK x blocks and the 4 PANC_BLK y blocks, results in the 16 possible coordinates of the trajectories that were learned via Lf D to be stored in robot's outputs. The signal that the learning has been completed (uE_Signaled) is generated by the lPANC_BLK network and makes it such that the procedures for moving the metallic workpieces are activated.

Imitation Step
After the information in the outputs (µ (k+1) ) of the lPANC_BLK function blocks indicate that the learning by demonstration of the gx and gy trajectories has been completed, the machine tool control system is activated and is responsible for activating the mobile device with a magnetic grabber.
With the 4 lPANC_BLK x blocks that store the transition states in the x axis and the 4 lPANC_BLK y blocks that store transition states in the y axis, it is possible to configure a linear Cartesian robot actuation with 16 different trajectories.

Results and Discussion
In this work, the lPANCells were arranged in a cascade forming two groups that compose the Paraconsistent Artificial Neural Network (lPANnet). This type of configuration allows 16 different trajectories to be obtained, where the transition states are represented by the values (X n ,Y n ) of the outputs of the lPANCells that were programmed in the form of the functional lPANC_BLK x and lPANC_BLK y blocks.
For this study, the 16 possible trajectories were classified into four types considering their linearity:

Learning Stage Results-Demonstration by Simulation
Initially, the structure that forms the lPANnet shown in Figure 9 was simulated to obtain the possible trajectories that could be learned by the linear Cartesian robot that acts on the machine tool. In the simulations of the lPANnet x and lPANnet y groups, the values of the patterns were applied directly to the input of the first cell programmed in the lPANC_BLK 1 function block. Initially, the pattern µ 1 = 0 was applied until all outputs indicated zero (µ (k+1) = 0). Then, after applying 11 iterations of the unit pattern (µ 1 = 1), the final condition of completed learning was considered. Table 4 contains the values obtained in the learning process through simulation, and the outputs of all lPANC_BLK x and lPANC_BLK y blocks that make up the lPANnet are shown in Figure 9. Table 4. Values obtained through simulation of the outputs from the lPANC_BLK x and lPANC_BLK y blocks that make up the lPANCnet that are shown in Figure 9.

Simulation Results-x-Axis (X 0 )-lPANC_BLK x
Simulation Results-y-Axis (Y 0 )-lPANC_BLK y These values from Table 4, which were obtained by simulation, will be compared to the values obtained in practice, as shown below.

Learning Stage Results-Demonstration by Teleoperation
In the procedures performed for learning, the blocks were initially activated by teleoperation with a pattern of 0 in the first lPANCell until all outputs resulted in zero (µ (k+1) = 0). Then, through the button (joystick), the control system was manually activated n times to move the sliding crossbar cylinder along the length of the table in the x axis until the "learning completed" signal.
The patterns values applied to the input of the first cell programmed in lPANC_BLK 1x function block were obtained from the x-axis transducer. After the learning procedure for the x-axis, the cylinder of the mobile device movement with a magnetic grip on the y-axis was manually activated through the button (joystick) until the "learning completed" signal occurred. The patterns values applied to the input of the first cell programmed in lPANC_BLK 1y function block were obtained from the y-axis transducer. Table 5 contains the values obtained in the learning process through teleoperation, and in the outputs of all lPANC_BLK x and lPANC_BLK y blocks that make up the lPANnet shown in Figure 9. Table 5. Values obtained through teleoperation in the outputs of the lPANC_BLK x and lPANC_BLK y blocks that make up the lPANnet shown in Figure 9.
It can be seen in Figure 10 that, although the trajectories a, b, c, and d are linear, there are fundamental differences in the distances of their transition states. This indicates that each path has a different velocity. Figure 11 shows the graphical results of the learning stage of the two different nonlinear trajectories that have a higher level of slope. The trajectories learned by the linear Cartesian robot of the machine tool are represented by the transition states obtained in the outputs of the cells (Figure 9): (µ E(k24)x , µ E(k11)y ) and (µ E(k21)x , µ E(k14)y ). Figure 12 shows the graphic results of the learning stage of the four different non-linear trajectories that have a medium level of slope. The trajectories learnt by the linear Cartesian robot of the machine tool are represented by the transition states obtained in the outputs of the following cells ( Figure 9): (µ E(k23)x , µ E(k11)y ), (µ E(k22)x , µ E(k14)y ), (µ E(k24)x , µ E(k12)y ), and (µ E(k21)x , µ E(k13)y ).
Note, in Figure 12, that although the non-linear trajectories b and d are similar, there are fundamental differences in the distances of their transition states. This indicates that each trajectory has different velocities and accelerations. Figure 13 shows the graphic results of the learning stage of the six different non-linear trajectories that have a low level of slope. The trajectories learned by the linear Cartesian robot of the machine tool are represented by the transition states obtained in the outputs of the following cells ( Figure 9): (µ E(k22)x , µ E(k11)y ), (µ E(k22)x , µ E(k13)y ), (µ E(k24)x , µ E(k13)y ), (µ E(k21)x , µ E(k12)y ), (µ E(k23)x , µ E(k12)y ), and (µ E(k23)x , µ E(k14)y ).
Note, in Figure 13, that the nonlinear trajectories exhibit fundamental differences in the distances of their transition states. This indicates that each trajectory has different velocities and accelerations. It can be seen in Figure 10 that, although the trajectories a, b, c, and d are linear, there are fundamental differences in the distances of their transition states. This indicates that each path has a different velocity. Figure 11 shows the graphical results of the learning stage of the two different nonlinear trajectories that have a higher level of slope. The trajectories learned by the linear Cartesian robot of the machine tool are represented by the transition states obtained in the outputs of the cells (Figure 9): (μE(k24)x, μE(k11)y) and (μE(k21) x, μE(k14)y).   , μE(k11)y), (μE(k22)x, μE(k14)y), (μE(k24)x, μE(k12)y), and (μE(k21) x, μE(k13)y).
Note, in Figure 12, that although the non-linear trajectories b and d are similar, there are fundamental differences in the distances of their transition states. This indicates that each trajectory has different velocities and accelerations.

Imitation Step-Results
After the teleoperation learning stage, the linear Cartesian robot receives the signal (uE_Signaled) warning that the imitation action is available. In this imitation stage, the x and y axis transducers allow real-time monitoring. Therefore, in the operation of the machine tool moving the workpieces, the transducers provide information about the realtime trajectory that is being developed by the linear Cartesian robot. Figure 14 shows the graphic results of the imitation stage that was performed by the linear Cartesian robot in a test on the gpn1 (µ E(k24)x , µ E(k11)y ) trajectory, which was classified as the one with the highest level of nonlinearity. Figure 15 shows the practical results compared to the simulated results.
As the graphs show in Figures 10-13, the trajectories not only differ in linearity levels, but are also different in their relation to the distances between the transition states. This variation between the distances of the transition states in the trajectories indicates that the parameters of velocity and acceleration are intrinsic to the results. However, in this work, where the tests were performed in a linear Cartesian robot driven by pneumatic energy, these parameters were not possible to analyze.

Imitation Step-Results
After the teleoperation learning stage, the linear Cartesian robot receives the signal (uE_Signaled) warning that the imitation action is available. In this imitation stage, the x and y axis transducers allow real-time monitoring. Therefore, in the operation of the machine tool moving the workpieces, the transducers provide information about the realtime trajectory that is being developed by the linear Cartesian robot. and y axis transducers allow real-time monitoring. Therefore, in the operation of the machine tool moving the workpieces, the transducers provide information about the realtime trajectory that is being developed by the linear Cartesian robot. Figure 14 shows the graphic results of the imitation stage that was performed by the linear Cartesian robot in a test on the gpn1 (μE(k24)x, μE(k11)y) trajectory, which was classified as the one with the highest level of nonlinearity.  The graphic results of the teleoperation learning stage showed that the linear Cartesian robot presented behaviors that were very close to the ideal results obtained by the simulation.
The graphical results shown in Figures 14 and 15 indicate that, when learning the trajectory via teleoperation, a recognition problem occurred in the first transition states. This problem, presented in the learning stage, was reflected in the imitation stage. In the example (Figure 15b), we can see that in the third transition state the simulated and practical values are as follows: Simulated Trajectory = (µ E(k21)x , µ E(k14)y ) = (0.75, 0.1875); Practical Trajectory = (µ E(k21)x , µ E(k14)y ) = (0.75, 0.3090).
When considering the simulated trajectory to be ideal, there is a maximum error of 64.8% in the practical trajectory toward the y axis. However, it appears that, both in the learning and imitation stage, the linear Cartesian robot showed that it has the conditions required to recover the ideal trajectory; this shows a robustness in the Lf D process to disturbances that are carried out by the lPANCells being implemented in this type of configuration.

Conclusions
In this article, we present a way of applying Paraconsistent Logic to the method of learning by demonstration (Lf D). Research on the Lf D method has been developed with the objective of making a machine perform new tasks by imitating procedures that are presented to it, without the need for reconfiguration or for a reprogramming of its software. Paraconsistent Logic (PL) is non-classical logic that is capable of processing data with contradictory information. As shown in this work, algorithms that are built based on Paraconsistent Annotated Logic (PAL), which is an extension of PL, can open up promising paths in this area of knowledge. In this research, we used the Paraconsistent Artificial Neural Cell of Learning (lPANCell) algorithm, which responds well to learning normalized patterns that have values between 0 and 1 being applied repeatedly to its input function. Thus, through recurrence techniques, lPANCell is able to gradually store this information, thus presenting signals with asymptotic variation as an output response, which are controlled by an adjustment of a learning factor (l F ). For the application of the Lf D technique with PAL, an implementation of the lPANCell algorithm was made in a Programmable Controller (PC) in the form of a functional block called lPANC_BLK, which was available in a virtual library according to the IEC 61131-3 standard. The modularity of the lPANC_BLK functional block guarantees that it can be used in the formation of configurations that form Paraconsistent Artificial Neural Networks (lPANnet), which are used in industrial automation projects that involve Programmable Controllers. The practical results were obtained with several tests being carried out to validate the operation of a lPANCell operating in a network structure composed of 8 lPANCells that were acting in a learning from demonstration (Lf D) process. The Lf D method, with the proposed lPANCnet configuration, was tested in a linear Cartesian robot, which worked as a machine tool for moving metallic workpieces. The lPANnet paraconsistent structure dedicated to "Learning" and "Imitation" was divided into two sets of four functional blocks each: where one acts in the learning process of linear Cartesian robot on the x-axis, and the other acts in the learning process of linear Cartesian robot on the y axis. The Lf D technique presented in this work was the learning by teleoperation variant, where, at the end of the procedures, transition states (Xn, Yn) for 16 trajectories of different linearity levels were created. The results obtained in the various tests carried out showed that the lPANnet configuration built with lPANC_BLK blocks is very efficient and responds adequately to all steps of the Lf D process. This indicates that the lPANnet paraconsistent structure is able to provide learning dynamic properties; further, it has a robustness to disturbances that can be monitored in real time, both in the learning process by demonstration and in the imitation process. Based on the results obtained in this work, new studies will be conducted by applying other different configurations with paraconsistent neural networks (lPANnet) in the Lf D process, thus covering more complex dynamic systems that are used in industrial plants and in the areas of robotics.