Abstract

Accurate event detection has high priority in many technical applications. Events in acquired data series, their duration, and statistical parameters provide useful information about the observed system and about its current state. This information can be used for condition monitoring, state identification, and many kinds of forecasting as well. In some cases background noise covers the events and simple threshold or power monitoring methods cannot be used effectively. A novel method called Scaled Sequential Probability Ratio Test (SSPRT) produces 2D array of data via special cumulative sum calculation. A peak determination algorithm has also been developed to find significant peaks and to store the corresponding data for further evaluation. The method provides straight information about the endpoints and possible duration of the detected events as well as shows their significance level. The new method also gives representative visual information about the structure of detected events. Application example for thermomechanical fatigue test monitoring and another for vibration based rotational speed estimation of a four-cylinder internal combustion engine is discussed in this paper.

1. Introduction

Since the century there has been a growing interest in statistical hypothesis testing. In the literature, several theories have been already proposed to extend the applicability of the mathematical background or to optimize calculation. However, new methodologies [14] and applications [510] of statistical hypothesis tests are published every year. The so-called Bayes’ theorem in probability theory has been established by Thomas Bayes in the early 1700s. Jerzy Neyman and Egon Pearson analysed the efficiency of hypothesis tests and have published their work in 1933 [11]. The Neyman-Pearson lemma offers a rule of thumb when all the data have been already collected. The lemma states that when performing a hypothesis test between two simple hypotheses and , the likelihood ratio test which rejects in favour of when where is the most powerful test at significance level for a threshold . The results of Neyman and Pearson inspired Abraham Wald in the mid-1940s to reformulate it as a sequential analysis problem [12]. A test can solely be called sequential, if the number of observations is not predetermined, but it is dependent on the outcome of the observations. A great summary about the evolution of sequential hypotheses tests and the early stages of the development can be found in Walds paper [12] in Chapter B.

In the last decades, there have been a surge of practical applications of the SPRT methodology in many areas including low frequency sonar detection, passive acoustic detection of marine mammals, tracking of signals, target tracking, early detection of changes in signals, computer simulations, data mining, clinical trials, gene ordering, agricultural sciences, horticulture, pest management, educational testing, economics and finance [13]. The latest published applications of SPRT focus for example on human core temperature prediction to prevent hyperthermia [5], diagnostics of the COMPASS tokamak [6], and W7X stellarator [10]; furthermore, many vibration measurement based applications of the methodology can be found in the literature [7, 8]. Some practical technical applications of SPRT have been also published by the authors. In these cases SPRT has been used for malfunction detection [14], place identification [15], or AE (acoustic emission) event detection [16]. However in the case of the referenced fusion diagnostic [6, 10] and AE applications [16] the duration of events seems to be a critically important information. Based on this information burst events derived from the test material could be separated from other types of events. Also in the case of vibration signal analysis of an internal combustion engine, setting limits for the length of detected events can help to make determination of cylinder activities more accurate. This paper discusses an AE application where ultrasonic burst events can be localised in noisy environment using the new methodology. The rate of hits shows clear correlation to the applied pulling force to the test specimen. Tachometerless speed estimation possibility of an internal combustion engine is also demonstrated.

In the following we introduce a novel extension to the classic SPRT method which produces a probability space and can help to get straight information about the most probable duration of detected events. In this paper a suggested peak detection method is also introduced which has been developed especially to evaluate the data resulted by the presented test method.

2. Scaled Sequential Probability Ratio Test

Introducing the Scaled Sequential Probability Ratio Test (SSPRT), it is an extended version of the classic SPRT method, which provides straight information about the endpoint and the duration of subsections of a signal with changed statistical distributions. In the following we refer to these changed subsections as events. The discussed methodology can be applied on discrete data series and the output data also stays discrete in time.

The methodology of the SSPRT can be separated into three main steps. First the likelihood ratio is calculated point by point via SPRT. Second a series of fixed size cumulative sums (cumulative sum array, CSA) are calculated sequentially with different predefined sizes. The range and the desired resolution of the observed event duration must be defined before the calculation starts. This way the method produces a 1D numeric array as output for every single numeric input value. After concatenating these 1D arrays the result changes into a 2D numeric array. This partial result of the SSPRT is quiet representative when displayed in a colormap, waterfall diagram, contour map, or as a 3D surface. The resulting surface shows a probability space, where the peaks determines the main properties (endpoint location and event duration) of the most possible event types. The third step of the SSPRT is the localisation of these local maximum points and selection of the most relevant ones.

2.1. Calculation of the Sequential Probability Ratio Test

This paper just gives a general overview about the background of SPRT. For definitions and detailed discussion of the theory please see [12]. We consider a simple hypothesis against a simple alternative . The SPRT for testing against is given as where denotes the observation on . According to the original method two constants are chosen where and both are dependent on the size of the acceptable decision errors. The errors are specified by and where and are small constants. The Cumulative Sum (CUSUM) of is calculated based on Then the rule of decision is made as follows:(i)if take another observation,(ii)if reject ,(iii)if accept .

The threshold values and and the error probabilities and are connected in the following way:Although Wald has shown that the SPRT procedure will stop with a decision with a finite number of observations, in some practical cases it is necessary to define a maximum test length. The procedure is then called the Truncated SPRT (TSPRT) [4, 7]. By truncating if the sequential process does not lead to a final decision after a given number of observation, an additional rule is given for the acceptance of hypothesis:(i)If , accept .(ii)If , accept .

However in this manner we change the probabilities of and . In the literature applications of fixed sample size tests can be also found. Other interesting extensions of the classical method are for multihypothesis testing [3, 17].

The first step of the SSPRT data evaluation is a special SPRT evaluation method, where parallel calculation of several SPRTs has to be done. The main idea of the algorithm is that if the value has already been calculated in the iteration, it can be reused for multiple hypothesis testing. The only criteria are that the different SPRT processes must be based on same hypothesis and the only difference between them is in the fixed run length. In our application at this first stage of calculation the lower and upper decision levels have been left out of consideration. In this manner the evaluation process always continues and never stops or restarts. This is because the SSPRT was developed for real-time applications and must be insensitive for initial value problems. For evaluation of the SPRT with fixed size log likelihood ratio (LLR) summation, a suitable method is the later discussed CUSUM algorithm. To make the final decision, one can easily use the rules of TSPRT already discussed before or define alternative decision methods and apply them in a later step.

During the tests of the implemented methods we used test signals mostly with normal distribution and zero mean. These terms makes the SPRT equations much simpler. The point-by-point formula is shown by (5):We can take some more simplifications if we introduce a new variable which marks the prior knowledge about the ratio of the deviations after and before the change point.where is the standard deviation of the whole signal and is the standard deviation of the changed subsections. If there are rare and relatively short events in the signal, we can say that is the standard deviation of the background noise, because of the negligible effect of the rare events. If we reorder the equation for , we get . Equation (7) shows the reordered form of (5).where and ) are constants and can be signed as and respectively. These constants can be computed and predefined. Equation (8) is the final equation for .where is the sample in the source data series. The parameter can be a predefined value or a sequentially computed parameter which can be easily refreshed using the observation. This way the SPRT will be an adaptive hypothesis test, which can change its sensitivity according to the background noise level. This implementation needs a few learning samples at the beginning of any evaluation to get the output with acceptable reliability.

2.2. Calculation of the Cumulative Sum Array

Many kind of sequential change point detection methods are available in the literature. These are usually called control charts or process-behaviour charts, but they are also known as Shewhart charts. Such elementary algorithms are for example the Combined Adaptive (CA) chart, Fixed Sampling Distance (FSD) chart, Fixed Sample Size (FSS) chart, Variable Sampling Distance (VSD) chart, Variable Sample Size (VSS) chart, Cumulative Sum (CUSUM) chart, Finite Moving Average (FMA) chart, and the Exponentially Weighted Moving Average (EWMA) chart. Most of the algorithms are used in practice work on samples of data with fixed size window or moving fixed size window.

The idea of the LLR-based CUSUM algorithm is that the prechange mean of the LLR is negative and the postchange mean is positive. If the analysed signal contains a changed subsection with a given length, the local result of the CUSUM [16] will be the highest at the end of the changed subsection if the cumulative calculation has been started just before the first change. In the case of our application a special implementation of the CUSUM chart has been used. The SSPRT algorithm realizes parallel sequential hypothesis tests. In this case the same hypothesis must be supposed for parallel threads. Only the probabilities of and of parallel TSPRTs varies because of the difference between the fixed numbers of summed LLR values. This way the LLR value must be calculated only once for a given sample in the source data series. The same value can be used to perform the actual iteration of the fixed size rolling CUSUM calculations. Henceforth the evaluation of a given data series using parallel CUSUM algorithms with different parameters is noted as CSA (Cumulated Sum Array). The CSA can be calculated in several ways and it results in a numeric array for every single numeric input. The calculation methodology will be discussed in later sections.

The CUSUM step of the SSPRT slightly differs from the conventional way of calculation. Many improved CUSUM methods are available in the literature. An interesting sequential solution has been called one-sided process inspection scheme or the extended two-sided version of it introduced by Page [18] has already much less sensitivity to the initial score because of the applied adaptive threshold. The one-sided version computes the cumulative sum as and takes action if . The rule can be reformulated as a sequential problem as follows in (9). The method takes action after the observation if .This method is sometimes called CUSUM Algorithm as a Repeated SPRT [17]. However in the SSPRT we need the “repeated” property of such algorithms but do not need the adaptive threshold because of using the decision rules of the TSPRT methodology discussed before. The SSPRT is a sequential algorithm which analyses the input series and makes decisions point by point. It must be totally insensitive to initial value settings and should never reset the CUSUM procedure because of special rules.

During SSPRT we take parallel CUSUM procedures simultaneously. One of these parallel processes can also be introduced as a special fixed sample size elementary change point detection algorithm which works on samples of data with fixed size sliding window.

Many publications show novel methods for optimized change detection using CUSUM as can be seen in references [2, 19] or analyse work efficient parallel scan solutions to get CUSUM values [20]. As in the case of these papers, the focus of our work has been to develop a general framework that offers scalable performance and operates in limited memory and in real-time. We supposed in the previous section that the events buried in the background noise are short and rare enough to keep the rolling parameter unbiased. We also suppose that the minimum and maximum length of the significant events are known and signed as and respectively. The SSPRT use a sliding window over the data stream. Its length is the amount of historical data that an algorithm stores and considers during the computation. The SSPRT method contains several parallel CUSUM processes, where only the length of the sliding window is different. The numeric results of these CUSUMs are the rows of the SSPRT output data. They operate with window sizes, where is a user defined parameter and notes the number of rows. The parameters must be chosen and . If a single number from the source data stream at time instant is and is the length of the sliding window, then the algorithm considers only the data points . As new data become available, the sliding window slides by one element, discarding the oldest and incorporating the latest data element. Now we show how to compute one row of the SSPRT if the parameter has been already chosen. If we follow a naive way to implement this algorithm, the equation will be as This algorithm needs addition operation for every single , i.e., to get one column of the SSPRT at time instant ; addition operations have to be done. This method needs to store the most recent pieces of input data in a buffer. The advantages of this algorithm are the simplicity and that using this it is not necessary to calculate all of the values in predefined order. One can easily skip unnecessary points or can also calculate the output values as independent processes via multicore processor system. The disadvantage is that this algorithm needs in normal continuous operation a very large number of addition operations and the number of operation depends on the chosen parameters. Another bad property is that the same data samples must be read from the buffer more times during the same iteration.

It must be noted that for example the corresponding subsections of the source data series of and have huge ( samples) overlap. If it is wanted to fulfil much less operation steps, one can easily reuse the outputs of the previous iteration. Now we show the recursive equation for calculation in (11).This implementation contains only one subtraction and one addition, i.e., to get one column of the SSPRT only numeric operations have to be done contrary to the previous algorithm. However it requires almost the same buffer size to store input data temporarily, and it also needs another buffer for the outputs of the previous iteration. It means one buffer of size and another of size .

We can take similar simplification if we take notice of the overlap of the corresponding subsection of and . Since , can also be computed as follows:In this case the overall process needs the same buffer size for the input elements. The advantage of this algorithm is that it does not need to store an array of previous output data. It uses only one output data from the previous CUSUM row instead.

In the next session two suggested scale types and the corresponding equations are shown. First of all we have to fix some basic parameters which define the properties of the chosen scale. These parameters are the minimum and maximum length of the sliding window expressed in number of samples ( and respectively) and the number of scales (). All of these parameters are positive integers and must be greater than zero: and .

2.2.1. Linear Scaled CSA

First we show a basic implementation where the parameter has to be ignored and . Now the only available setting is and the resolution of the output will be this way the highest because for .

Using one of the above described CSA equations we get an output array for every single input data, and the sequence evaluates the presence of subsections that have been changed in the input data series and have got the given length. The resolution of the observed length is this way the best, equaling the sampling rate of the input data series. When using the default , both of the CSA II and the CSA III algorithms can be used easily. If an alternative is required, the CSA II seems to be practical, because in the case of the CSA III the rows that correspond to window sizes below must be also computed to get the final result. These additional output values can be released after the actual column has been completely calculated.

The duration of the detected events are often less important as the accurate localisation. To decrease the buffer sizes and to get an output via less overlapped calculations, a possible way is choosing less than the observed range of length.

A modified parameter can be defined as , where . This is the feasible resolution according to the chosen parameters. The slight modification of the parameter also modifies the feasible range () of the analysis as . First, we have to compute (13) for the optimal distance between two windows that are neighbours.The array of the length of all used windows can be now expressed as Let see an example to show how this scaling method works. Let the basic parameters be , , m=8. The input data is generated as a series of negative and positive units. The negative units illustrate the LLR which corresponds to the background noise and the positive values show where the statistic has been changed. Two events are simulated in this example. The first one starts at the second sample and has only two samples. The other one starts at the fifth sample of the series and has four samples. It can happen that the classification of some samples, which has been acquired during an event, is incorrect. If two or more detected short events are near enough to each other, that they also can be parts of a more dominant long change must be analysed.

Figure 1 shows the output data array (), where the bottom row () equals to the input data series because . In the figure the white cells indicate negative data points. According to the decision rules of the TSPRT method, in the case of grey or black cells the alternative hypothesis is accepted; i.e., the signal has been changed. The SSPRT can be a special decision procedure, which results in point-by-point decisions corresponding to the most recent input samples. On the other hand the output can be seen as a special probability space, where local peaks show the most probable parameter combinations (location and duration) of a possible event.

As it can be seen in the presented example, two changed subsections are in the input data series. Two of the three highlighted local peaks are at the same location. It means that the last four samples are parts of a changed subsection, but it is also possible that these points are parts of a longer event which contains seven data points before the time instant. In a later chapter it will be discussed how is it possible to locate the most relevant peaks in this data array in a sequential way.

2.2.2. Logarithmic Scaled CSA

In the case of many applications logarithmic scaling makes the result much informative. In the case of the lower section of the logarithmic scale, short changes can be observed much better because of the better resolution.

Let us introduce a new parameter which is quite similar to the parameter in the case of linear scales. Here shows the ratio of the window length of two neighbour scales and . If the number of requested CSA rows () and are also predefined input parameters, one have to find the optimal value:Based on the latter (16) defines the array for a logarithmic scale.

2.3. Peak Detection

Several mathematical methods are available to detect peaks in data series. It is a difficult question, which one is the best choice to use in our application. We chose a specialised method. In the case of the SSPRT it is a special property that the neighbour rows in the output are results of overlapped sum calculation with different window sizes. Because of the applied rectangular windows another important property is that all rows are scaled, delayed, and low-pass filtered versions of the original input data. The presented 2D output data of SSPRT can be also computed as an output array of a special low-pass filtering filter bank. CSA I and CSA III, which have been presented in a previous section of this paper, represent different ways of calculation to get the same result. These two methods follows the same logic of calculation as techniques for FIR (Finite Impulse Response) and IIR (Infinite Impulse Response) filtering respectively. The CSA II method is also similar to FIR techniques as CSA I. The former represents just an optimized way to get the same output.

Another important question may arise if one develops a peak detection method for 2D data arrays. Do we need all local peaks? In our case there is a time instant where two peaks can be found in the same column. The value of these peaks represents the probability of the corresponding subsections to be an event, while unchanged sections of the input (contains only background noise) result in negative LLRs and events (changed subsections) result in positive LLRs. If a peak has higher maximum value than another, the applied window has longer overlap to the detected event. In such cases the longer event with higher peak value has to be stored and lower peaks in the same column can be ignored.

In the case of our example series illustrated by Figure 1, three local maximum points can be found. The coordinates of these points in the three-dimensional Descartes coordinate system are , , . Based on the latter considerations just two of the peaks are relevant and . These points correspond to overlapping subsections of the original signal so we should decide which one to store as a final decision. Similarly to the case of overlapped subsections which have been detected at the same time instance, now we can do the same comparison and decision procedure for the applied window length and the peak magnitude. Based on probability ratio test the final decision is that the signal contains a changed subsection which starts at the time instance and has a length of seven samples.

Now let us see some mathematical equations and how this logic has been implemented as a sequential evaluation method. In the following the methodology is discussed in four main steps.

Step 1. At first step, the maximum points of the sequentially calculated columns of CSA have to be found. At all time instances, the maximum values are searched and the coordinates of the resulting maximum point are stored in a new array. These parameter sets (time instance or column index; row index; maximum value of the analysed column) are noted as .In the series the coordinates of the maximum point and the corresponding value can be stored. However it is practical to drop the row index and use the corresponding sliding window size instead. This parameter describes the calculation process better and can be useful in later steps if already has been stored. During sequential event detection methods the length of an event can only be determined when the event has already ended and the used window size is a key parameter at this point if the start date of the event () is needed.

Step 2. The decision logic has to be sequential. It means we only know the values before and at the examined time instance. First, possible local maximum points in the series of the cumulated LLR values have to be determined. If the current value is greater than or equal to the value of the previous iteration, a partial decision can be stated that the point can be a local maximum (LM) point of the series.where is the Boolean result of the decision about the input data which could be a local maximum. Boolean outputs as 0 and 1 represent false and true values respectively.
Second, the evaluation of the applied SPRT rules should be an alert if the analysed subset of the data series could correspond to an event.where is the result of the local decision of the applied SPRT logic. The threshold parameter discussed earlier at the introduction of SPRT can be configured to zero if one follows the evaluation methodology of TSPRT but can also be chosen to any positive value, which affects the accuracy and the sensitivity of the classical SPRT methodology.
If both partial decisions result true, the actual data cluster has to be stored for later use. The as the holding decision at time instance can be determined as .
is a new series, which holds according to the decision. can be determined point by point using the following equations:

Step 3. The goal of this step is to find the endpoint of the detected events. Now the decision result at time instance will be signed as .Any time if the results true, it can be stated that a detected event or statistical change just ended. This is not the final decision yet, because these alerts can also sign tiny event parts, which can also be a part of a long change. It must be cleared that the possible correspondence to a longer event can only be determined later.

Step 4. In this step the previously detected possible events are evaluated. These events can correspond to the same event or they can be registered as individual events. To determine this property, a watchdog logic has been used. The watchdog must restart its operation if the actual becomes true. The initial value when restart is needed equals always the actual value. The watchdog () decrements its stored value by one if false and alerts if the stored value reaches the zero level. This alert is signed now as which means the final decision at time instance .Once a final decision result has been determined as true, the actual property values stored in correspond to the most possible event detected in the last section of the source signal. These results are signed as as event properties after final decision at time instance .In the case of the example series shown in Figure 1, the final decision about the detected event parameters is as follows:This means that a decision could be made at the time instance, that a statistical change has been occurred in the source signal starting at the time instance, and its most probable duration is seven samples.

3. Using the SSPRT Method for Analysing Generated Test Signal

The applicability of the developed SSPRT methodology has been tested on several types of test signals. The method seemed to be working as expected in the case of generated test signals and on real life measurement signals as well. The following data series and figures are just representative examples for the evaluation of a generated test signal. Detailed discussion of applications will be topics for later publication.

The following example in Figure 2 shows a tiny section of the generated signal that contains Gaussian white noise with unit RMS (Root Mean Square) as constant background and periodically contains additional Gaussian noise with the same power and 0.2 seconds duration. The sections with additional noise are the simulated events which have to be detected using the SSPRT methodology. The additional noise results slightly in increased power of the signal between 10 and 10.2 seconds, and it still have Gaussian distribution during the event. If one takes a look at the generated time signal which is shown by Figure 2, it can be stated for the first sight that a common threshold test cannot be used for reliable event detection and also the start and end points of the events cannot be determined accurately. If the common SPRT evaluation method has been chosen we get the data series as Lambda values shown by Figure 3. The common SPRT is able to detect the simulated event as the decision function has reached the level of 6.9 more times. It can be seen that event detection works well, if LLR values have been processed using the common evaluation method, but how to get exact duration of the event remains a question. In the case of the novel SSPRT evaluation method we need to apply one of the above suggested CSA calculation methods to the LLR values. The fixed size cumulative sum calculation results in the highest value, if the length of the applied window fits to the event duration and if these two have full overlap. It can be easily demonstrated. In the case of a too short window, not all event-related positive LLR values are used for sum calculation, and too long windows occur that background noise related negative LLR values are taken into account.

Figure 4 illustrates the resulting probability surface where the highest peak shows the most probable duration end endpoint of the detected event. Determined duration of the statistical change has been 192 ms and its probable start time 9.999 s. These are extremely good estimations for the 200 ms long simulated event beginning at 10 s.

Local maximum points of the resulting three-dimensional surface can be easily found by several peak searching algorithms which can be applied to the presented data of the above colormap. The previously described mathematical background gives a possible solution to realize an effective peak searching method especially for the CSA output.

4. Practical Applications of SSPRT

SSPRT can be a versatile tool for any applications, where changes of the noisy input signal have to be detected accurately and the length of events is a key parameter for further evaluation. In the case of any practical applications where common SPRT evaluation has been applied successfully on sampled data series, the novel SSPRT methodology will provide at least that accurate event detection. Furthermore SSPRT shows directly the most probable duration which can help in event classification or source identification.

In the case of AE applications only well optimized sequential algorithms can be used for real-time evaluation. Steel under heat or pressure stresses emits AE signals, which can be detected using AE sensors. Experiments have been planned and carried out using a material testing simulator shown in Figure 5. Detailed description of the experimental setup and the PXIe-6363 based data acquisition system can be found in reference [16, 21]. The results showed that the SPRT method is absolutely suitable for AE event detection during fatigue test. Figure 6 clearly demonstrates the relationship between the applied loading force and the rate of detected AE events. The test specimen produced much more ultrasonic bursts during pulling forces than in the case of compression.

One of the detected bursts can be seen in Figure 7 and the corresponding CSA result of the SSPRT evaluation in Figure 8. In the case of the evaluation methodology introduced by reference [16] smoothing was applied to the LLR values before locating start and end time of events. It seems to work well, but this way averaging parameters must be fixed before evaluation. The disadvantage of the moving average method is that rapid internal changes during an event cannot be observed; this is also mentioned in that article. SSPRT could be a better tool for analysing such events’ internal structure. If we take a look at the colormap, the highest peak shows that the detected event is 237.8μs long, starts at 264.9μs, and lasts for 502.7μs. The resulting data can be also displayed as a three-dimensional surface, which provide new opportunities for further studies. Other AE applications are for example failure-monitoring of slide bearings [22] or localisation of initial cracks in windscreen glass [23], where SSPRT could be a great tool for accurate event detection, localisation, and analysis.

Based on the above, SSPRT is well suited for detection of rare events in noisy environment. This is guaranteed by the hypothesis test theory, and it is possible to calculate the probability of correct and incorrect decisions. If events occur more frequently than expected by the theory or if the distribution does not match the hypotheses, the SSPRT will still be able to detect events. It remains also possible to analyse internal structure of events. The disadvantage of periodically appearing changes is that decision making can be done at a lower significance level. By SSPRT evaluation of vibration acceleration of an internal combustion engine, it is easy to detect combustion processes of the cylinders. If the SSPRT method can reliably detect all cylinder activities, further conclusions can be drawn. For example combustion frequency is proportional to the engines rotational speed. By displaying and processing the resulting three-dimensional data, detailed information can be obtained about the performance of combustion processes in each cylinder. Knowing the exact position of the crankshaft gives the possibility of resampling signals by angular rotation and then cyclostationary diagnostic tools can be easily applied. The results of such algorithms can be displayed in polar diagrams, which offer further possibilities of measurement evaluation.

In Figure 9 the raw vibration acceleration data series and the derived results appear simultaneously. Event lengths resulted by SSPRT are illustrated as transparent vertical bars. Depending on the event length, different colored bars appear in the graph: short events are green (2 to 5 ms); yellow bars sign medium length (6 to 7 ms); the longest ones appear in red (8 ms). In the (c) graph the determined combustion frequency can be seen. This frequency can be easily converted to rotational speed. The experimental setup contained a high precision encoder as reference sensor. Average speed of the gas motor was 1455 rpm which equals the 48.5 Hz average combustion frequency resulted by SSPRT.

5. Conclusions

This article presents a new technique, which enhances the evaluation method of the classical sequential probability ratio test. This novel method is based on realization of parallel fixed size cumulative sum calculations with different predefined sizes on the calculated log likelihood ratio. The resulting data is called cumulative sum array and the values can be visualized as a three-dimensional surface. This paper has shown three alternative ways, how the same cumulative sum array values can be calculated. Some special advantages and disadvantages of these alternative ways of the calculation have been also discussed. Many peak detection methods could be used to evaluate the resulting data of the cumulative sum array; however, the authors suggest specialised peak detection algorithm. It has been developed to evaluate the log likelihood ratio based cumulative sum array sequentially. If a local peak has been found, the coordinates of the peak show the queried properties of the event. Rows correspond to a given event length, columns show when an event has been ended. Maximum values of the peaks give information about strength and significance. The proposed method can be readily used in practice. The successful applicability of the novel evaluation method has been first demonstrated on generated test signals where the noisy background contained short sections with additional noise. Thereafter, practical examples have been presented through acoustic emission and vibration diagnostic applications. The method can also be useful by cumulated sum array visualization on color maps. Graphical representation makes it much easier to compare detected events via giving a good visual overlook of the events’ internal structure.

Data Availability

The data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare that there are no conflicts of interest regarding the publication of this paper.

Acknowledgments

The research reported in this paper was supported by the Higher Education Excellence Program of the Ministry of Human Capacities in the frame of Artificial Intelligence research area of Budapest University of Technology and Economics (BME FIKP-MI/FM).