Next Article in Journal
A Simple Wireless Sensor Node System for Electricity Monitoring Applications: Design, Integration, and Testing with Different Piezoelectric Energy Harvesters
Next Article in Special Issue
Dynamic Speed Control of Unmanned Aerial Vehicles for Data Collection under Internet of Things
Previous Article in Journal
Wheat Height Estimation Using LiDAR in Comparison to Ultrasonic Sensor and UAS
Previous Article in Special Issue
Development of Sensor Registry System-Based Predictive Information Service Using a Grid
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Adaptive Parallel Processing Strategy for Complex Event Processing Systems over Data Streams in Wireless Sensor Networks

1
School of Computer and Information Science, Southwest University, No.2 Tiansheng Road, BeiBei District, Chongqing 400715, China
2
Big Data Science and Technology, Division of Informatics and Energy, Faculty of Advanced Science and Technology, Kumamoto University, 2-39-1 Kurokami, Chuo-ku, Kumamoto 860-8555, Japan
*
Authors to whom correspondence should be addressed.
Sensors 2018, 18(11), 3732; https://doi.org/10.3390/s18113732
Submission received: 23 September 2018 / Revised: 22 October 2018 / Accepted: 30 October 2018 / Published: 2 November 2018
(This article belongs to the Special Issue Internet of Things and Ubiquitous Sensing)

Abstract

:
Efficient matching of incoming events of data streams to persistent queries is fundamental to event stream processing systems in wireless sensor networks. These applications require dealing with high volume and continuous data streams with fast processing time on distributed complex event processing (CEP) systems. Therefore, a well-managed parallel processing technique is needed for improving the performance of the system. However, the specific properties of pattern operators in the CEP systems increase the difficulties of the parallel processing problem. To address these issues, a parallelization model and an adaptive parallel processing strategy are proposed for the complex event processing by introducing a histogram and utilizing the probability and queue theory. The proposed strategy can estimate the optimal event splitting policy, which can suit the most recent workload conditions such that the selected policy has the least expected waiting time for further processing of the arriving events. The proposed strategy can keep the CEP system running fast under the variation of the time window sizes of operators and the input rates of streams. Finally, the utility of our work is demonstrated through the experiments on the StreamBase system.

1. Introduction

Recently, there has been an increasing interest in wireless sensor networks, which require continuously processing flowing data from geographically-distributed sources to achieve timely responses to complex queries, such as data stream processing (DSP) systems [1,2,3] and complex event processing (CEP) systems [4,5,6]. Additionally, the CEP systems focus on detecting the patterns of information that represent the higher level events, which are different with the DSP systems that focus on transforming the incoming flow of information [7]. Because the CEP system has many advantages, such as an expressive rule language and an efficient detection model of events, it has been highly concerned in academic circles and recently in industry [8,9,10,11,12,13,14]. In the CEP systems over data streams, events are processed in real time for all kinds of purposes, such as wireless sensor networks, financial tickers, traffic management, click-stream inspection and smart hospitals [15,16,17]. In these application domains, highly available event stream processing with a fast processing time is critical for handling the real-world events.
As far as we know, many kinds of parallel methods were devised to deal with massive distributed data streams for the DSP systems [18,19,20,21,22,23,24,25,26]. However, due to the differences between the DSP and CEP systems, most of the parallel methods that exclusively focus on aggregate queries or binary equi-joins in the DSP systems cannot be simply and directly used in the CEP systems that focus on multi-relational non-equi-joins on the time dimension, possibly with temporal ordering constraints, such as the sequence (SEQ) operator and conjunction (AND) operator [27,28]. Furthermore, the large volume and input rates of data streams are very common in big data applications [29,30]. The increased time window sizes of operators and input rates of streams may cause bottlenecks for the CEP system. Bottlenecks can slow down the CEP system. Even worse, they can result in the poor quality of query results, which has negative effects on decision-making.
To address these issues, we propose a parallelization model and an adaptive parallel processing strategy, called APPS by introducing histogram, probability theory and queue theory. The proposed APPS can estimate the optimal event splitting policy, which suits the most recent workload conditions such that the selected policy has the least expected waiting time for further processing of the coming events. Specifically, the CEP system based on the proposed parallelization model can split the input stream into parallel sub-streams to realize a scalable execution of continuous pattern query. APPS can keep the CEP system operating at high speed even under the variation of the time window sizes of the operators and input rates of the streams. The utility of our work is substantiated through the experiments on the StreamBase [12] system.
The rest of this paper is organized as follows. Section 2 discusses the related work in terms of the CEP systems. Section 3 briefly introduces the preliminaries of this paper. After that, a parallelization model and three event splitting policies are proposed in Section 4. Then, an adaptive parallel processing strategy is proposed to estimate and select the optimal event splitting policy to suit the workload conditions in Section 5. Section 6 demonstrates the utility of our proposal through the experiments on the StreamBase system. Finally, conclusions are given in Section 7.

2. Related Work

In CEP systems, the operators are demanded to be highly scalable under high event stream rates. It is well known that the traditional CEP systems are mostly centralized. For increasing the scalability of CEP systems, distributed parallelization processing is necessary. Until recently, studies were conducted and were mainly classified into two types: one focuses on the task parallelization of CEP systems; the other focuses on the data parallelization of CEP systems.
The task parallelization is also known as pipelining or intra-operator parallelization, where by deriving the states of operators and state transitions from the pattern query, the internal processing steps can be identified to be run in parallel. In particular, Suhothayan et al. [31] brought multi-threading and pipelining into CEP systems to make them process quickly. Wu et al. [32] presented a framework of parallelization for stateful operators over stream processing. Although these task parallelization methods are effective, they are not feasible for pattern matching. Later on, Balkesen et al. [33] devised a parallelization within a single partition of an event stream for scalable pattern matching. For the data parallelization, the main research works are as follows. Brito et al. [34] presented a system by combining the responsiveness of event stream processing systems with the scalability of the MapReduceprogramming model. Schneider et al. [35] introduced a compiler and run-time system, which could automatically extract data parallelism from streaming applications by partitioning the state through keys. De Matteis and Mencagli [36] presented parallel patterns for window-based stateful operators.
Through analyzing the existing works, it was found that the presented task parallelization and data parallelization methods were limited by the function of operators, especially for the pattern operators. Besides, some existing works that were based on key splitting were limited to the number of different key values. Most importantly, there is no adaptive parallel processing strategy for pattern operators in CEP systems. Therefore, in this paper, we focus on adaptive parallelization of pattern operators in CEP systems, which is a main contribution of this study. It is clear that this work contrasts withand is complementary to the previous works.

3. Preliminaries

3.1. Event Model

An event that represents an instance and is atomic is an occurrence of interest at a point in time. Basically, events can be classified into primitive events and composite events. A primitive event instance is a pre-defined single occurrence of interest that cannot be split into any small events. A composite event instance that occurs over an interval is created by composing primitive events.
Definition 1.
A primitive event e i is typically modeled multi-dimensionally, denoted as e i = e( e i . t , ( e i . s t = e i . e t ), < a 1 , …, a m > ), where, for simplicity, we use the subscript i attached to a primitive e to denote the timestamp i, e i . t is the event type that describes the essential features of e i , e i . s t is the start time-stamp of e i , e i . e t is the end time-stamp of e i , < a 1 , …, a m > are other attributes of e i and the number of attributes in e ( · ) denotes the dimensions of interest.
Definition 2.
Based on Definition 1, a composite event is denoted as e=e( e . t , (( e . s t = min 1 i n e i . s t ) < ( e . e t = max 1 i n e i . e t )), < a 1 , …, a g > ).

3.2. Nested Pattern Query Language

We introduce the following nested complex event query language for specifying nested pattern queries:
  • PATTERN (event expression: composite event expressed by the nesting of SEQ and AND, which can have negative event type(s), and their combination operators)
  • WHERE (qualification: value constraint)
  • WITHIN (window: time constraint)
The composite event expression in the PATTERN clause specifies nested pattern queries, which support nests of SEQ and AND that can have negative event type(s), and their combination operators, as explained above. The negative event type in PATTERN means that the composite event is generated only when this event type has not occurred. Sub-expressions denote inner parts of a pattern query expression. The value constraint in the WHERE clause defines the context for the composite events by imposing predicates on event attributes. The time constraint in the WITHIN clause describes the time window during the time difference between the first and the last event instances, which is matched by a pattern query that falls within the window.

3.3. Pattern Operators and Their Formal Semantics

We define the operators that our method is targeting. Specifically, in this paper, we consider the pattern operators as presented in [37,38]. In the following, E i denotes an event type. More details were presented in [28].
Definition 3.
An SEQ operator [37,38] specifies a particular order according to the start time-stamps in which the event must occur to match the pattern and thus form a composite event:
S E Q ( E i , E j ) = { < e i , e j > | ( e i . s t < e j . s t ) ( e i . t = E i ) ( e j . t = E j ) } .
Definition 4.
An AND operator [38] takes the event types as input, and events occur within a specified time window without a specified time order:
A N D ( E i , E j ) = { < e i , e j > | ( e i . t = E i ) ( e j . t = E j ) } .

4. System Model

4.1. Parallelization Model

In this section, we propose a parallelization model that can be utilized for pattern operators, which is shown in Figure 1. We assume that each pattern operator is installed into a server (or host) here. Because of the specific property of pattern operators as described in Section 3, we cannot split both inputs I E i and I E j at the same time. Otherwise, this will omit detecting some events that may result in the wrong decision. Without loss of generality, an input stream can be randomly selected to be split, while the other one is replicated. Here, we assume splitting the input stream I E j and replicate the other stream I E i . Specifically, once an event of I E j arrives, the compute function of the pattern operator is initiated. In other words, the pattern operator creates a new window for every input tuple of I E j . Therefore, the input stream I E j is split into parallel sub-streams that will be sent to back-end operators. The input rate of stream I E j is equal to the sum of the input rates of sub-streams, i.e., λ E j = k = 1 m λ E j , k , where λ E j , k represents the input rate of a sub-stream to the back-end operator k. On the other hand, the replicate of input stream I E i is directly sent to the back-end operators, each of which has input rate λ E i . We now provide details of the s p l i t ( p r o c e s s ) m e r g e assembly, which facilitates the parallelization model of pattern operators.
As shown in Figure 1, the s p l i t ( p r o c e s s ) m e r g e assembly replaces the solo pattern operator in the application data-flow. In the parallelized version of the application data-flow, λ E j is split to the back-end process operators, and the output of the pattern operator is replaced by the output coming from the merge operator.
  • Split: The split operator is to split an input stream into parallel sub-streams. The split operator outputs the incoming events to a number of back-end pattern operators by one of the event splitting policies from Section 4.2, where this selected event splitting policy is estimated by the adaptive parallel processing strategy that will be explained in Section 5.
  • Process: The process operator performs the events from the output of the front-end operators. The multiple process operators with the same function can be executed in parallel.
  • Merge: The merge operator consumes the output events from the process operators to generate the final output events. The merge operator by default simply forwards the output events to its output port.

4.2. Event Splitting Policies

In this section, the event splitting policies are given, which can be utilized for processing pattern operators in parallel.
• Round-robin (RR):
Events are assigned to the servers in a cyclical fashion, which means that the incoming events will be sent to the downstream servers with equal probability. This policy equalizes the expected number of events at each server.
• Join-the-shortest-queue (JSQ):
For the expected number of events, they are assigned to the downstream server with the shortest queue length for further processing. Here, the shortest queue means the queue with the fewest events.
• Least-loaded-server-first (LLSF):
For the expected number of events, it dynamically assigns them to the downstream server with the least load. The least loaded server is the server with the least used memory.

5. Adaptive Parallel Processing Strategy

In this section, an adaptive parallel processing strategy (APPS) is proposed to estimate and select the optimal event splitting policy, which can suit the most recent workload conditions such that the selected policy has the least expected waiting time for processing the coming events. Table 1 shows the key notations that are used in the remainder of this paper. Figure 2 describes the flowchart of the adaptive parallel processing strategy.

5.1. Degrees of Parallelization

The aim of this stage is to decide the degrees of parallelization for pattern operators in the CEP system to be used for processing data streams.
Let ρ be the expected server utilization, μ be the service rate, m be the number of servers and δ be the threshold of the expected server utilization that can be defined by the system administrator in advance according to the implication requirement. By applying queueing theory [39], ρ is given by:
ρ = λ m μ , s . t . ρ δ , 0 < δ 1 .
Based on Equation (1), we can obtain the degrees of parallelization for the pattern operator, i.e., the number of processing servers is as follows:
m λ μ δ , 0 < δ 1 ; m N .

5.2. Expected Size of the Batch Partition

For further parallel processing, the input stream I E j needs to be divided into batch partitions. Because the number of events within each batch partition of segment S ν of input stream I E j should not exceed the threshold of the expected utilization of a single server, the number of events of a batch partition i should satisfy the following condition:
i = μ δ , i N .

5.3. Event Processing Time Collection

The aim of this stage is to collect the processing time of the events from the last event type matched by the pattern operator, which are used in the on-line estimation step to estimate various distributional properties of the processing time distribution.
For each new event arriving at the split operator, it records the arrival time of the event. These values are stored within the event. The arriving events of a segment of the input stream are then assigned to a back-end server by using the estimated policy P j , where P j denotes the event splitting policy with the least expected waiting time to process the arriving events. Further details about how to select an appropriate event splitting policy on-line are discussed in Section 5.5. For each event that completes processing, its departure time will be stored at the assigned server. Next, the arrival time and departure time of the event will be sent to its back-end operator. Then, its corresponding processing time will be calculated by subtracting the departure time, which contributes to the last output event matched by the pattern operator that falls within the time window, from the arrival time recorded by the split operator.

5.4. Trade-Off between the Estimation Accuracy and the Processing Time

Figure 3 depicts an example of obtaining an appropriate policy for processing the further coming events. S ω denotes the ω th segment of input stream I E j . B 1 represents the first batch partition of the segment, which consists of events { e 1 , e 2 , , e i } . The policy P j under segment S ν means these events in S ν will be processed by using P j in which this estimated P j is selected based on the empirical data S ω . Therefore, we can notice that the time devoted to processing previous number of segments over m parallel servers should exceed the time devoted to estimating an appropriate policy P j for segment S ν . Otherwise, it introduces extra delay due to the waiting to obtain the optimal policy. In addition, to obtain the most accurate expected processing time for S ν , the mean squared error is considered.
Consequently, we treat the accuracy of estimation and the processing time of segments as an integrated constrained optimization problem. One objective ( O 1 ) tries to maximize the accuracy of estimation, namely minimizing the mean squared error of estimation. On the other hand, the other objective ( O 2 ) tries to maximize the processing time of segments to avoid introducing extra delay in selecting the optimal policy. Due to the conflicting nature of the different objectives, we obtain the solution by integrating them into one objective, and the optimization problem can thus be formulated as:
min O 1 O 2 .
On the basis of the statement (4), the values of q and can be obtained and be used in the further on-line event splitting policy selection procedure in Section 5.5.
Objective O 1 : mean squared error of estimation constraint.
In this paper, a general expression is derived for the expected waiting time by applying queueing theory [40], denoted as f ( E [ W ] ) . Let f ^ S ω ( E [ W ] ) be the expected processing time of the events of segment S ν in terms of empirical data S ω . Then, f ^ S ω ( E [ W ] ) is compared with f S ( ν 1 ) ( E [ W ] ) by the following mean squared error ( M S E ):
M S E = 1 τ q 1 ω = ν 1 , ν = + 2 τ q ( f ^ S ω ( E [ W ] ) f S ( ν 1 ) ( E [ W ] ) ) 2 , w i t h f ( E [ W ] ) = ρ 2 ( m + 1 ) 1 μ m ( 1 ρ ) ( C a 2 + C s 2 2 ) , 1 q , τ ; 1 ω τ q , s . t . M S E < β ,
in which q is the number of batch partitions of segment S ν and is the difference value subtracting ( ν 1 ) of S ν 1 from ω of S ω , which is devoted to estimating policy P j for segment S ν . C a 2 represents the squared coefficient of variation of inter-arrival times, and C s 2 represents the squared coefficient of variation of service times where they can be obtained by testing. f S ( ν 1 ) ( E [ W ] ) is the true expected processing time of the events of segment S ν in terms of empirical data S ( ν 1 ) , because it is the nearest empirical data of S ν to obtain the most accurate expected processing time. β denotes the threshold of the mean squared error of estimation that can be defined by the system administrator in advance according to the implication requirement.
Objective O 2 : processing time constraint.
Let T ¯ p s i be the time devoted to processing i number of events, T ¯ r d i + 1 be the time devoted to re-directing the ( i + 1 ) th event and T ¯ e s i be the estimate time for i number of events. The objectives O 2 should satisfy the following condition:
T ¯ p s S = q T ¯ p s i + ( q 1 ) T ¯ r d i + 1 , s . t . T ¯ p s S m > T e s P j , w i t h T e s P j = q T ¯ e s i , 1 q , τ .
The values of T ¯ p s i , T ¯ r d i + 1 and T ¯ e s i can be obtained via testing. If the number of events within one batch partition of S ν is large enough, while the time for re-directing each batch partition is quite smaller than the time for processing each batch partition, we can omit T ¯ r d i + 1 in Equation (6).

5.5. On-Line Selection of Event Splitting Policies

This stage is pretty critical in the proposed adaptive parallel processing strategy, which can estimate and decide the appropriate policy on-line. In order to calculate the expected waiting time for the policies, we first leverage the histogram to obtain the probabilities that the events are sent to host i by policy P j , denoted as P j H i , and the probabilities that the events are redirected to host i by policy P j , denoted as P j R i .
Next, we introduce queue theory [41] to get the expected waiting time for the events at host i, which is formulated as:
E [ W i H ] = 1 μ i ( ρ i 1 ρ i ) ( C i a 2 + C i s 2 2 ) ,
where ρ i denotes the expected server utilization at host i, μ i represents the number of events served per unit time at host i, C i a 2 represents the squared coefficient of variation of inter-arrival times at host i and C i s 2 represents the squared coefficient of variation of service times at host i where they can be obtained on-line.
Additionally, we utilize probability theory to calculate the expected redirect time for the events at host i, which is formulated as:
E [ W i R ] = r = 1 k x r f ( x r ) .
Based on the probabilities that events are sent and redirected to different hosts, the expected waiting time for the events at different hosts and the expected redirect time at different hosts, we then calculate the expected waiting time for all the policies in the list of APPS. APPS derives a general expression for the expected waiting time for policy P j , denoted as E [ W P j ] , by applying probability theory to select the event splitting policy with the least expected waiting time, which can be formulated as:
E [ W P j ] = i = 1 h ( P j H i E [ W i H ] + P j R i E [ W i R ] ) .

6. Experimental Evaluation

Based on the parallelization model in Figure 1, we implemented the experiments on the StreamBase [12] system for query q 1 .
q 1 : PATTERN S E Q ( E 1 , E 2 ) WHERE [ Id ] WITHIN 1 s
Since the proposed method both contrasts withand is complementary to the existing methods, APPS is compared with the RR, JSQ and LLSF methods to prove the utility and effectiveness of the proposed method. We ran the experiments on the machines, each of which has an AMD Opteron(tm) Processor 6376 and 4.00 GB main memory. Streams used in the experiments were generated synthetically. Specifically, each stream was set with three attributes, including the event id, time-stamp and event type, in which the incoming events of streams were uniformly distributed. We define the processing time as the difference between the departure time, which contributes to the last output event matched by the pattern operator that falls within the time window, and the arrival time recorded by the split operator. For the simplicity of the experiments, we provided four machines for APPS, RR, JSQ and LLSF: one machine that created input data and split the input stream into back-end machines, another two machines that were equipped with S E Q operators with the same functions to process the input streams in parallel and another machine that received data and output the throughput. Then, we compared the performance of these methods under different parameter settings in terms of input rate and time window size.

6.1. Comparing the Processing Time of the Methods

In this experiment, the input rates were set as 100 events/s, and time window sizes were set as 1 s. From the experimental result as shown in Figure 4, it was obvious that APPS and JSQ had lower processing time compared with the RR and LLSF methods. Because APPS could estimate and select the optimal event splitting policy for further processing of the coming events, it had almost the same processing time as the JSQ method.

6.2. Varying the Time Window Sizes of Operators

In this experiment, the input rates were set as 100 events/s, while the time window sizes were varying from 1 up to 10, and 100 s. From the experimental result as shown in Figure 5, we can notice that APPS had almost the same processing time as the JSQ method, and both outperformed the RR and LLSF methods, especially when the time windows sizes were set as 1 s; whereas, as the time windows sizes increased from 10 up to 100 s, the APPS, RR, JSQ and LLSF methods almost had the same performance. The reason is that as the time windows sizes increased from 10 up to 100 s, it reached the limitation of the processing capacity of the machines.

6.3. Varying the Input Rates of Streams

In this experiment, the time window sizes were set as 1 s, while the input rates were varying from 100 up to 200, 300 and 400 events/s. From the experimental result as shown in Figure 6, we can obviously see that the performance of APPS was significantly better than the performance of the RR, JSQ and LLSF methods. Because APPS, which suits the most recent workload conditions, estimated and selected the optimal event splitting policy for further processing of the coming events, it could handle the input rate variation environment. On the other hand, as the input rates increased from 100 up to 200, 300 and 400 events/s, the JSQ and LLSF methods had a lower processing time than the RR method, because JSQ assigned the events to the back-end server with the shortest queue length, while LLSF assigned the events to the back-end server with the least load for further processing of the coming events.

7. Conclusions

In this paper, we started off with identifying the general problems of adaptive parallel processing with respect to pattern operators in CEP systems. We proposed a new adaptive parallel processing strategy to estimate the optimal event splitting policy, which can suit the most recent workload conditions such that the selected policy had the least expected waiting time for further processing the coming events. Moreover, because the proposed method was a hybrid method of intra-operator parallelization and data parallelization, our proposal was not limited to the number of different key values. The proposed strategy kept the CEP system running with fast processing under the variation of the time window sizes of operators and input rates of streams. The utility of our work was demonstrated through the experiments on the StreamBase system.
The proposed adaptive parallel processing strategy considered only the SEQ and AND operator-based pattern query in this study. Thereby, in the future work, more complex operators will be further investigated for the pattern queries, such as the nested SEQ and AND operators, which may have negative event types, and combinations of them. In addition, we intend to achieve a pilot implementation of the framework, where more complicated experimental environment and performance analysis will be taken into account in the future work, including the Poisson distribution, the exponential distribution of incoming events, etc. [42,43]. Another interesting future work is about how to detect complex events over probabilistic event streams adaptively. In view of the efficiency in handling uncertainty, some useful extended methods, like fuzzy theory, evidence theory, probability and the entropy-based method [44,45,46,47], will be considered in CEP systems in the future work.

Author Contributions

F.X. designed the method, performed the experiments and wrote the paper. M.A. designed the method, examined the experiments and revised the paper.

Funding

This research is supported by the JSPS KAKENHI Grant (No. 15H02705), the Chongqing Overseas Scholars Innovation Program (No. cx2018077), the National Natural Science Foundation of China (Nos. 61672435, 61702427, 61702426) and the 1000-Plan of Chongqing by Southwest University (No. SWU116007).

Acknowledgments

The author greatly appreciates the reviews’ suggestions and the editor’s encouragement.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lee, I.; Park, J.H. A scalable and adaptive video streaming framework over multiple paths. Multimed. Tools Appl. 2010, 47, 207–224. [Google Scholar] [CrossRef]
  2. Ding, J.W.; Deng, D.J.; Lo, Y.K.; Park, J.H. Perceptual quality based error control for scalable on-demand streaming in next-generation wireless networks. Telecommun. Syst. 2013, 52, 445–459. [Google Scholar] [CrossRef]
  3. Jang, J.; Jung, I.Y.; Park, J.H. An effective handling of secure data stream in IoT. Appl. Soft Comput. 2018, 68, 811–820. [Google Scholar] [CrossRef]
  4. Chen, M.Y.; Wu, M.N.; Chen, C.C.; Chen, Y.L.; Lin, H.E. Recommendation-aware smartphone sensing system. J. Appl. Res. Technol. 2014, 12, 1040–1050. [Google Scholar] [CrossRef]
  5. Boubeta-Puig, J.; Ortiz, G.; Medina-Bulo, I. A model-driven approach for facilitating user-friendly design of complex event patterns. Expert Syst. Appl. 2014, 41, 445–456. [Google Scholar] [CrossRef] [Green Version]
  6. Macià, H.; Valero, V.; Díaz, G.; Boubeta-Puig, J.; Ortiz, G. Complex event processing modeling by prioritized colored Petri nets. IEEE Access 2016, 4, 7425–7439. [Google Scholar] [CrossRef]
  7. Cugola, G.; Margara, A. Processing flows of information: From data stream to complex event processing. ACM Comput. Surv. (CSUR) 2012, 44, 15. [Google Scholar] [CrossRef]
  8. SASE. Available online: http://avid.cs.umass.edu/sase/ (accessed on 12 September 2017).
  9. Cayuga. Available online: http://www.cs.cornell.edu/bigreddata/cayuga/ (accessed on 12 September 2017).
  10. PIPES. Available online: http://dbs.mathematik.uni-marburg.de/Home/Research/Projects/PIPES/ (accessed on 12 September 2017).
  11. Coral8. Available online: http://www.complexevents.com/coral8-inc/ (accessed on 12 September 2017).
  12. Streambase. Available online: https://www.tibco.com/products/tibco-streambase (accessed on 12 September 2017).
  13. Oracle CEP. Available online: https://www.oracle.com/technetwork/middleware/complex-event-processing/overview/index.html (accessed on 12 September 2017).
  14. CEP for Hospital. Available online: https://stanfordhealthcare.org/search-results.clinics.html (accessed on 12 September 2017).
  15. Boubeta-Puig, J.; Ortiz, G.; Medina-Bulo, I. ModeL4CEP: Graphical domain-specific modeling languages for CEP domains and event patterns. Expert Syst. Appl. 2015, 42, 8095–8110. [Google Scholar] [CrossRef]
  16. Kim, K.; Kim, H.; Kim, S.K.; Jung, J.Y. i-RM: An intelligent risk management framework for context-aware ubiquitous cold chain logistics. Expert Syst. Appl. 2016, 46, 463–473. [Google Scholar] [CrossRef]
  17. Xiao, F.; Aritsugi, M.; Wang, Q.; Zhang, R. Efficient processing of multiple nested event pattern queries over multi-dimensional event streams based on a triaxial hierarchical model. Artif. Intell. Med. 2016, 72, 56–71. [Google Scholar] [CrossRef] [PubMed]
  18. Safaei, A.A.; Haghjoo, M.S. Parallel processing of continuous queries over data streams. Distrib. Parallel Databases 2010, 28, 93–118. [Google Scholar] [CrossRef]
  19. Han, W.S.; Kwak, W.; Lee, J.; Lohman, G.M.; Markl, V. Parallelizing query optimization. Proc. VLDB Endow. 2008, 1, 188–200. [Google Scholar] [CrossRef] [Green Version]
  20. Hirzel, M. Partition and compose: Parallel complex event processing. In Proceedings of the 6th ACM International Conference on Distributed Event-Based Systems, Berlin, Germany, 16–20 July 2012; pp. 191–200. [Google Scholar]
  21. Johnson, T.; Muthukrishnan, M.S.; Shkapenyuk, V.; Spatscheck, O. Query-aware partitioning for monitoring massive network data streams. In Proceedings of the 24th International Conference on Data Engineering, Vancouver, BC, Canada, 9–12 June 2008; pp. 1135–1146. [Google Scholar]
  22. Liu, B.; Rundensteiner, E.A. Revisiting pipelined parallelism in multi-join query processing. In Proceedings of the 31st International Conference on Very Large Data Bases, Trondheim, Norway, 30 August–2 September 2005; pp. 829–840. [Google Scholar]
  23. Chaiken, R.; Jenkins, B.; Larson, P.Å.; Ramsey, B.; Shakib, D.; Weaver, S.; Zhou, J. SCOPE: Easy and efficient parallel processing of massive data sets. Proc. VLDB Endow. 2008, 1, 1265–1276. [Google Scholar] [CrossRef]
  24. Upadhyaya, P.; Kwon, Y.; Balazinska, M. A latency and fault-tolerance optimizer for online parallel query plans. In Proceedings of the ACM SIGMOD International Conference on Management of Data, Athens, Greece, 12–16 June 2011; pp. 241–252. [Google Scholar]
  25. Safaei, A.A.; Haghjoo, M.S. Dispatching stream operators in parallel execution of continuous queries. J. Supercomput. 2012, 61, 619–641. [Google Scholar] [CrossRef]
  26. Brenna, L.; Gehrke, J.; Hong, M.; Johansen, D. Distributed event stream processing with non-deterministic finite automata. In Proceedings of the Third ACM International Conference on Distributed Event-Based Systems, Nashville, TN, USA, 6–9 July 2009; p. 3. [Google Scholar]
  27. Akdere, M.; Çetintemel, U.; Tatbul, N. Plan-based complex event detection across distributed sources. Proc. VLDB Endow. 2008, 1, 66–77. [Google Scholar] [CrossRef] [Green Version]
  28. Xiao, F.; Aritsugi, M. Nested pattern queries processing optimization over multi-dimensional event streams. In Proceedings of the 37th Annual Computer Software and Applications Conference, Kyoto, Japan, 22–26 July 2013; pp. 74–83. [Google Scholar]
  29. Carney, D.; Çetintemel, U.; Cherniack, M.; Convey, C.; Lee, S.; Seidman, G.; Stonebraker, M.; Tatbul, N.; Zdonik, S. Monitoring streams: A new class of data management applications. In Proceedings of the 28th International Conference on Very Large Data Bases, Hong Kong, China, 20–23August 2002; pp. 215–226. [Google Scholar]
  30. Xiao, F.; Kitasuka, T.; Aritsugi, M. Economical and fault-tolerant load balancing in distributed stream processing systems. IEICE Trans. Inf. Syst. 2012, 95, 1062–1073. [Google Scholar] [CrossRef]
  31. Suhothayan, S.; Gajasinghe, K.; Loku Narangoda, I.; Chaturanga, S.; Perera, S.; Nanayakkara, V. Siddhi: A second look at complex event processing architectures. In Proceedings of the 2011 ACM Workshop on Gateway Computing Environments, Seattle, WA, USA, 18 November 2011; pp. 43–50. [Google Scholar]
  32. Wu, S.; Kumar, V.; Wu, K.L.; Ooi, B.C. Parallelizing stateful operators in a distributed stream processing system: How, should you and how much? In Proceedings of the 6th ACM International Conference on Distributed Event-Based Systems, Berlin, Germany, 16–20 July 2012; pp. 278–289. [Google Scholar]
  33. Balkesen, C.; Dindar, N.; Wetter, M.; Tatbul, N. RIP: Run-based intra-query parallelism for scalable complex event processing. In Proceedings of the 7th ACM International Conference on Distributed Event-based Systems, Arlington, TX, USA, 29 June–3 July 2013; pp. 3–14. [Google Scholar]
  34. Brito, A.; Martin, A.; Knauth, T.; Creutz, S.; Becker, D.; Weigert, S.; Fetzer, C. Scalable and low-latency data processing with stream mapreduce. In Proceedings of the IEEE Third International Conference on Cloud Computing Technology and Science (CloudCom), Athens, Greece, 29 November–1 December 2011; pp. 48–58. [Google Scholar]
  35. Schneider, S.; Hirzel, M.; Gedik, B.; Wu, K.L. Auto-parallelizing stateful distributed streaming applications. In Proceedings of the 21st international conference on Parallel Architectures and Compilation Techniques, Minneapolis, MN, USA, 19–23 September 2012; pp. 53–64. [Google Scholar]
  36. De Matteis, T.; Mencagli, G. Parallel patterns for window-based stateful operators on data streams: An algorithmic skeleton approach. Int. J. Parallel Program. 2017, 45, 382–401. [Google Scholar] [CrossRef]
  37. Liu, M.; Rundensteiner, E.; Greenfield, K.; Gupta, C.; Wang, S.; Ari, I.; Mehta, A. E-Cube: Multi-dimensional event sequence analysis using hierarchical pattern query sharing. In Proceedings of the ACM SIGMOD International Conference on Management of Data, Athens, Greece, 12–16 June 2011; pp. 889–900. [Google Scholar]
  38. Liu, M.; Rundensteiner, E.; Dougherty, D.; Gupta, C.; Wang, S.; Ari, I.; Mehta, A. High-performance nested CEP query processing over event streams. In Proceedings of the 27th International Conference on Data Engineering, Hannover, Germany, 11–16 April 2011; pp. 123–134. [Google Scholar]
  39. Dattatreya, G.R. Performance Analysis of Queuing and Computer Networks; CRC Press: Boca Raton, FL, USA, 2008. [Google Scholar]
  40. Newell, C. Applications of Queueing Theory; Springer Science & Business Media: Berlin, Germany, 2013; Volume 4. [Google Scholar]
  41. Saaty, T.L. Elements of Queueing Theory: With Applications; McGraw-Hill: New York, NY, USA, 1961; Volume 34203. [Google Scholar]
  42. Ficco, M.; Esposito, C.; Palmieri, F.; Castiglione, A. A coral-reefs and game theory-based approach for optimizing elastic cloud resource allocation. Future Gen. Comput. Syst. 2018, 78, 343–352. [Google Scholar] [CrossRef]
  43. Ficco, M.; Pietrantuono, R.; Russo, S. Aging-related performance anomalies in the apache storm stream processing system. Future Gen. Comput. Syst. 2018, 86, 975–994. [Google Scholar] [CrossRef]
  44. Yin, L.; Deng, X.; Deng, Y. The negation of a basic probability assignment. IEEE Trans. Fuzzy Syst. 2018. [Google Scholar] [CrossRef]
  45. Kang, B.; Deng, Y.; Hewage, K.; Sadiq, R. Generating Z-number based on OWA weights using maximum entropy. Int. J. Intell. Syst. 2018, 33, 1745–1755. [Google Scholar] [CrossRef]
  46. Fei, L.; Deng, Y. A new divergence measure for basic probability assignment and its applications in extremely uncertain environments. Int. J. Intell. Syst. 2018. [Google Scholar] [CrossRef]
  47. Zhang, W.; Deng, Y. Combining conflicting evidence using the DEMATEL method. Soft Comput. 2018. [Google Scholar] [CrossRef]
Figure 1. The parallelization model.
Figure 1. The parallelization model.
Sensors 18 03732 g001
Figure 2. The flowchart of the adaptive parallel processing strategy.
Figure 2. The flowchart of the adaptive parallel processing strategy.
Sensors 18 03732 g002
Figure 3. Example of obtaining an appropriate policy for processing the coming events.
Figure 3. Example of obtaining an appropriate policy for processing the coming events.
Sensors 18 03732 g003
Figure 4. Comparing the processing time of the methods (APPS: An adaptive parallel processing strategy, RR: Round-robin, JSQ: Join-the-shortest-queue, LLSF: Least-loaded-server-first).
Figure 4. Comparing the processing time of the methods (APPS: An adaptive parallel processing strategy, RR: Round-robin, JSQ: Join-the-shortest-queue, LLSF: Least-loaded-server-first).
Sensors 18 03732 g004
Figure 5. Comparing the methods under the variation of time window sizes.
Figure 5. Comparing the methods under the variation of time window sizes.
Sensors 18 03732 g005
Figure 6. Comparing the methods under the variation of input rates.
Figure 6. Comparing the methods under the variation of input rates.
Sensors 18 03732 g006
Table 1. Notation.
Table 1. Notation.
NotationMeaning
P j event splitting policy j
ρ the expected server utilization
δ threshold of the expected server utilization
mdegree of parallelization of servers
μ number of events served per unit time
λ E j input rate of input stream I E j
S ν the ν th segment of input stream I E j
B g the g th batch partition of a segment
inumber of events of a batch partition
qnumber of batch partitions of a segment
T ¯ p s i average time devoted to processing i number of events
T ¯ r d i + 1 average time devoted to re-directing the ( i + 1 ) th event
among servers
T ¯ p s S average time devoted to processing segments
T ¯ e s i average estimation time devoted for i number of events
T e s P j estimation time devoted to obtaining optimal P j for S ν
E [ W i R ] expected redirect time for the events at host i
E [ W i H ] expected waiting time for the events at host i
E [ W P j ] expected waiting time for policy P j

Share and Cite

MDPI and ACS Style

Xiao, F.; Aritsugi, M. An Adaptive Parallel Processing Strategy for Complex Event Processing Systems over Data Streams in Wireless Sensor Networks. Sensors 2018, 18, 3732. https://doi.org/10.3390/s18113732

AMA Style

Xiao F, Aritsugi M. An Adaptive Parallel Processing Strategy for Complex Event Processing Systems over Data Streams in Wireless Sensor Networks. Sensors. 2018; 18(11):3732. https://doi.org/10.3390/s18113732

Chicago/Turabian Style

Xiao, Fuyuan, and Masayoshi Aritsugi. 2018. "An Adaptive Parallel Processing Strategy for Complex Event Processing Systems over Data Streams in Wireless Sensor Networks" Sensors 18, no. 11: 3732. https://doi.org/10.3390/s18113732

APA Style

Xiao, F., & Aritsugi, M. (2018). An Adaptive Parallel Processing Strategy for Complex Event Processing Systems over Data Streams in Wireless Sensor Networks. Sensors, 18(11), 3732. https://doi.org/10.3390/s18113732

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop