Abstract

In big data parallel processing, parallel defects, e.g., data race and deadlock, are common causes that affect reliability of programs. Uncertainty in parallel processing characterises parallel defects, for which fuzziness of time sequence analysis plays an important role. To improve the performance of big data processing, we propose a multithreaded time sequence analysis approach based on type-2 fuzzy logic and hidden Markov model in this paper. Firstly, we collect a sample set of training data by carrying out extensive experiments for the target multi-threaded program with given observations. Secondly, we establish a time sequence analysis model to describe the inner relationship between the observations and time sequence of the target multi-threaded program. Thirdly, using this model we estimate the probability of each state sequence in all the target defect positions, with which we estimate the probability of defects for the corresponding observation sequence. To prove the scalability in a big data environment, we also use our approach to analyse a real concurrency defect in real world large-scale multi-thread programs. Our experiment results show that the average deviation using type-2 fuzzy logic is less than one fourth of the average deviation using type-1 fuzzy logic.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call