1. Introduction
The exchange of goods and services is one of the activities that materialize human interactions. This quasi-routine activity involves physical entities or people, as well as legal entities such as businesses and corporations.
The volume of trade over the last ten years shows that the number of goods and services traded has steadily increased throughout history [1] . The size and number of exchanges are accompanied, among other things, by 1) difficulties in managing and forecasting the quantities to be ordered in order to obtain supplies, build up reserves, and ensure that stock does not run out or overflow; and 2) customer behavior uncertainty manifested by ignorance of the quantities and frequency with which they can buy over a given period of time. These last activities, carried out solely by human intuition on a daily basis, increase the likely margins of error in estimating salable and/or flow out quantities.
Marketing storable goods are activities whose realization recalls random events and/or processes that can be executed over time. As a result, Markov chains accurately model this type of process. They are distinguished by the fact that the probabilities implying how the process will evolve in the future are determined solely by the current state of the process and are thus independent of past events [2] . A stochastic process is a random phenomenon that evolves over time.
The Markov chain in stochastic analysis specifies a system of transitions of an entity from one state to another. A Markov chain’s states are divided into two categories: transient states, which are only visited an almost finite number of times (p.s.), and recurrent states, which are visited an almost infinite number of times (p.s.), as well as all other states in the same recurrence class. A recurrent irreducible Markov chain and the process’s marginal law converge to either the unique P-invariant probability measure or the null vector or null recurrence. This theory is particularly applicable to random markets and queuing models.
The need to ensure the reliability of a system arose around the turn of the twentieth century, peaking during World War II. Currently, the study of dependability is evolving into its own discipline. Economic issues, for example, necessitate a certain level of inherent dependability. The ability to predict system reliability is critical for ensuring the safety of braking systems, nuclear systems, computer systems, and other systems. In short, reliability analysis is a critical step in any study of a system’s dependability [3] .
It is critical to characterize a system’s behavior throughout its life cycle. The virtual impossibility of repairing certain equipment (satellites), the cost of system failures, spare parts inventories, and so on demonstrates the importance of understanding the reliability of the systems used. Because failures generally occur at random, it is logical to use probability calculations to strengthen their reliability through prediction. The reliability of a device, defined as its probability of functioning correctly for a given duration, is equivalent to determining the likelihood that no failure will occur during this duration [4] .
2. Materials, Tools, Equipment and Methods
Random events or processes of product supply management necessitate the use of appropriate materials and methods. As stated in the introduction, this work makes use of the tools provided by Markov chains and its eventual applications.
2.1. Material
Because we are dealing with processes for managing a system with discrete random events, the implementation of such a system is centered on the use of adjacency matrices and state transition matrices, as well as the design of process-specific execution algorithms. As stated in the introduction, this work makes use of the capabilities provided by Markov chains as materials and equipment.
2.1.1. Discrete-Time Markov Chains
Let
be a sequence of random variables with values in the set of states E assumed to be equal to N. We say that this sequence is a Markov chain if
and any sequence
, we have:
(1)
probability distributions characterize this chain:
1) One-dimensional distribution:
(2)
2) The conditional probability distribution (also known as the transition probability distribution):
(3)
Consider
to be a Markov chain with values in S.
The Markov chain is said to be homogeneous if the probability of passing to state j at time n + 1 knowing that we were in state i at time n does not depend on n:
for all
. (4)
where are, the transition probabilities of the Markov chain are denoted by the numbers
,
.
2.1.2. Probabilities and Transition Matrix
The probability that X(n) is in state
is one of the state probabilities
, the sum of which is obviously unity:
(5)
while
are the probabilities of one-step transitions from state i to state j at time n, satisfying the relation:
(6)
These transition probabilities are generally written in the form of a matrix called the Transition Matrix or Stochastic Matrix:
(7)
whose sum of the elements in the same row equals one.
The probabilities of two-step transitions from state i to state j are calculated as follows:
(8)
which is written as a matrix:
(9)
We can use the Chapman-Kolmogorov equations to obtain the m-step transition probabilities from state i to state j:
(10)
for all
;
;
;
.
These equations show that if you go from state i to state j in m steps, you will arrive at state k after exactly l (l being less than m) states. Thus,
is only the conditional probability that, given a starting point of state j, the process proceeds to state k after l steps, then to state i after m − l steps. As a result, the sum of these conditional probabilities over all possible k must equal
. The special case of l = 1 yields Equation (10):
(11)
or in a matrix form
(12)
2.2. Methods
Inventory and supply management is an important aspect of a business’s operations. When it is effective, it reduces overstocking and increases the overall profitability of the company. Every procurement manager is faced with the daily task of making the best use of their products storage, namely inventory and supply storage management. At daily basis, he should dynamically modify inventories without exposing the company to risk of stock-out [5] . Overstocks result in additional costs due to storage fees on the one hand, while stocks out result in a revenue loss due to the client’s selflessness in the company on the other. The Markov’s chain addresses the inventory and supply process with threshold limits using probabilistic laws and principles.
Thus, a careful examination of the events that occur during the marketing of storable goods (stock) reveals that these processes can be correctly represented or modeled by Markov chains. The temporary and temporal storage of these goods can be studied using Markov chain concepts. A Markov chain is a random process
, the transitions of which are determined by a stochastic matrix
. These processes demonstrate the Markov property, namely that observed from a (stopping) time T,
depends solely on XT and is, once again, a Markov chain. Here are some major algorithms related to Markov chain mathematical tool. Figures 1-4 show the implemented Markov chain algorithms, while Figure 5 shows the transition probability. The all algorithms presented in this sub-subsection are all written in the form of program in high level language Java.
2.2.1. Algorithm for the Markov Chain Test
The Markov chain testing algorithm is depicted in the Java source code in Figure 1. The final one ensures that the modeled processes exactly correspond to the Markov chain.
2.2.2. Algorithm for Calculating Matrices’ Products
As shown in Figure 2, one of the main methods in state transition matrices is the matrices’ products algorithm.
2.2.3. Chapman-Kolmogorov Equation Solution Algorithm
The algorithm for solving the Chapman-Kolmogorov equations computes the transition probabilities, which are then used to determine the stationary probability state. Figure 3 shows the algorithm as depicted in the java source code.
2.2.4. Algorithm for Constructing State Transition Matrices
The Transition matrix allows us to determine the stock state at any given time. Figure 4 depicts the algorithm in pseudo-code form using the Java programming language paradigm.
2.2.5. Function for Calculating Transition Probability
The transition probability function shows the probability transition from one state to another at a given time. Figure 5 depicts the function used to calculate transition probabilities during the prediction process.
![]()
Figure 1. Algorithm for the Markov chain test.
![]()
![]()
Figure 2. Algorithm for calculating matrices’ products [6] .
![]()
Figure 3. Chapman-Kolmogorov equation solution algorithm.
![]()
![]()
Figure 4. Algorithm for constructing state transition matrices.
![]()
Figure 5. Function for calculating transition probability.
3. Results and Discussion
3.1. Results
3.1.1. Algorithm for the Markov Chain Test
The Markov chain testing algorithm ensures that the processes being modeled exactly correspond to the Markov chain and thus have Markov properties.
3.1.2. Chapman-Kolmogorov Equation Solution Algorithm
The algorithm for solving the Chapman-Kolmogorov equations allows for the determination or calculation of transition probabilities, including those in the steady state.
3.1.3. Algorithm for Constructing State Transition Matrices
The transition matrix construction algorithm enables the creation of a “mirror” of all possible states for a system. The algorithm for calculating the transition probability, on the other hand, displays the state of a system at a given instant over a specified time period.
3.2. Discussions
Similar to the supply management processes and tools that can be built using Markov chains, guaranteeing the reliability of a technical system, particularly an artificial one, over a given period is equivalent to calculating that no failure will occur during this period. As a result, the theory of reliability, whose goal is to investigate the ability of technical devices to perform a required function under given conditions, during a given time, and for a given period, necessitates a mathematical modeling tool analogous to Markov chains. Satisfying the standards of reliability of technical systems through the application of probability theory of random phenomena appears to be a broad axis of application for the model designed and algorithms developed in this work.
In this era of increased commercial competition, decision-making by trial and error and/or improvised approximation based on natural human thought has only resulted in certain bankruptcy for the institution concerned. Thus, the algorithms developed in this article, when applied to automated inventory management models, provide numerous benefits, including reduced product storage costs and reduced shortfalls due to stock depletion.
This is demonstrated, for example, by the scientific contribution and applicability of the algorithm for solving the Chapman-Kolmogorov equations, the algorithm for constructing the transition matrix and the algorithm for calculating the probability of state transition.
4. Conclusions
The tools for implementing a model, no matter how well designed, are critical in ultimately constructing a management system with the desired technical specifications and obtaining the expected results.
The computer application of the discrete-time Markov chain model allows a company to manage the supply, storage, and sale processes of countable products without shortages and/or overstocks. Such a working method increases the system’s dependability.
Using an illustrative example, this work demonstrated that the predictability of the reliability of a system for managing supplies, stocks, and sales of a company selling storable goods allows for the limitation, if not elimination, of the phenomena of stock depletion and/or stock overflow, thanks to a simple and ergonomic management interface in which the algorithms developed will be implemented. These results show that the management and decision-making support model accurately predicts the quantity in stock as well as the quantity to be ordered, avoiding the phenomenon of storage products exhaustion or overflow.
As a result, this tool ensures that additional costs due to additional warehousing costs and potential losses due to out-of-stock situations are avoided. Finally, the implementation of simple and ergonomic management interface algorithms simplifies product storage monitoring procedures and makes managers’ daily tasks easier.