Towards Low-Latency Batched Stream Processing by Pre-Scheduling
Research output: Contribution to journal › Journal article › Research › peer-review
Standard
Towards Low-Latency Batched Stream Processing by Pre-Scheduling. / Jin, Hai; Chen, Fei; Wu, Song; Yao, Yin; Liu, Zhiyi; Gu, Lin; Zhou, Yongluan.
In: IEEE Transactions on Parallel and Distributed Systems, Vol. 30, No. 3, 8444732, 2019, p. 710-722.Research output: Contribution to journal › Journal article › Research › peer-review
Harvard
APA
Vancouver
Author
Bibtex
}
RIS
TY - JOUR
T1 - Towards Low-Latency Batched Stream Processing by Pre-Scheduling
AU - Jin, Hai
AU - Chen, Fei
AU - Wu, Song
AU - Yao, Yin
AU - Liu, Zhiyi
AU - Gu, Lin
AU - Zhou, Yongluan
PY - 2019
Y1 - 2019
N2 - Many stream processing frameworks have been developed to meet the requirements of real-time processing. Among them, batched stream processing frameworks are widely advocated with the consideration of their fault-tolerance and high throughput. In batched stream processing frameworks, straggler, happened due to the uneven task execution time, has been regarded as a major hurdle of latency-sensitive applications. Existing straggler mitigation techniques, operating in either reactive or proactive manner, are all post-scheduling methods, and therefore inevitably result in high resource overhead or long job completion time. We notice that batched stream processing jobs are usually recurring with predictable characteristics. By exploring such a feature, we present a pre-scheduling straggler mitigation framework called Lever. Lever first identifies potential stragglers and evaluates nodes capacity by analyzing execution information of historical jobs. Then, Lever carefully pre-schedules job input data to each node before task scheduling so as to mitigate potential stragglers. We implement Lever and contribute it as an extension of Apache Spark Streaming. Our experimental results show that Lever can reduce job completion time by 30.72% to 42.19% over Spark Streaming, a widely adopted batched stream processing system and outperforms traditional techniques significantly.
AB - Many stream processing frameworks have been developed to meet the requirements of real-time processing. Among them, batched stream processing frameworks are widely advocated with the consideration of their fault-tolerance and high throughput. In batched stream processing frameworks, straggler, happened due to the uneven task execution time, has been regarded as a major hurdle of latency-sensitive applications. Existing straggler mitigation techniques, operating in either reactive or proactive manner, are all post-scheduling methods, and therefore inevitably result in high resource overhead or long job completion time. We notice that batched stream processing jobs are usually recurring with predictable characteristics. By exploring such a feature, we present a pre-scheduling straggler mitigation framework called Lever. Lever first identifies potential stragglers and evaluates nodes capacity by analyzing execution information of historical jobs. Then, Lever carefully pre-schedules job input data to each node before task scheduling so as to mitigate potential stragglers. We implement Lever and contribute it as an extension of Apache Spark Streaming. Our experimental results show that Lever can reduce job completion time by 30.72% to 42.19% over Spark Streaming, a widely adopted batched stream processing system and outperforms traditional techniques significantly.
KW - data assignment
KW - recurring jobs
KW - scheduling
KW - straggler
KW - stream processing
UR - http://www.scopus.com/inward/record.url?scp=85052709429&partnerID=8YFLogxK
U2 - 10.1109/TPDS.2018.2866581
DO - 10.1109/TPDS.2018.2866581
M3 - Journal article
AN - SCOPUS:85052709429
VL - 30
SP - 710
EP - 722
JO - IEEE Transactions on Parallel and Distributed Systems
JF - IEEE Transactions on Parallel and Distributed Systems
SN - 1045-9219
IS - 3
M1 - 8444732
ER -
ID: 203670980