TY - CONF
T1 - The case for tiny tasks in compute clusters
AU - Ousterhout, Kay
AU - Panda, Aurojit
AU - Rosen, Joshua
AU - Venkataraman, Shivaram
AU - Xin, Reynold
AU - Ratnasamy, Sylvia
AU - Shenker, Scott
AU - Stoica, Ion
N1 - Funding Information:
We thank Matei Zaharia, Colin Scott, John Ouster-hout, and Patrick Wendell for useful feedback on earlier drafts of this paper. This research is supported in part by NSF CISE Expeditions award CCF-1139158 and DARPA XData Award FA8750-12-2-0331; gifts from Amazon Web Services, Google, SAP, Blue Goji, Cisco, Clearstory Data, Cloudera, Ericsson, Facebook, General Electric, Hortonworks, Huawei, Intel, Microsoft, Ne-tApp, Oracle, Quanta, Samsung, Splunk, VMware and Yahoo!; and a Hertz Foundation Fellowship.
PY - 2013
Y1 - 2013
N2 - We argue for breaking data-parallel jobs in compute clusters into tiny tasks that each complete in hundreds of milliseconds. Tiny tasks avoid the need for complex skew mitigation techniques: by breaking a large job into millions of tiny tasks, work will be evenly spread over available resources by the scheduler. Furthermore, tiny tasks alleviate long wait times seen in today's clusters for interactive jobs: even large batch jobs can be split into small tasks that finish quickly. We demonstrate a 5.2x improvement in response times due to the use of smaller tasks. In current data-parallel computing frameworks, high task launch overheads and scalability limitations prevent users from running short tasks. Recent research has addressed many of these bottlenecks; we discuss remaining challenges and propose a task execution framework that can efficiently support tiny tasks.
AB - We argue for breaking data-parallel jobs in compute clusters into tiny tasks that each complete in hundreds of milliseconds. Tiny tasks avoid the need for complex skew mitigation techniques: by breaking a large job into millions of tiny tasks, work will be evenly spread over available resources by the scheduler. Furthermore, tiny tasks alleviate long wait times seen in today's clusters for interactive jobs: even large batch jobs can be split into small tasks that finish quickly. We demonstrate a 5.2x improvement in response times due to the use of smaller tasks. In current data-parallel computing frameworks, high task launch overheads and scalability limitations prevent users from running short tasks. Recent research has addressed many of these bottlenecks; we discuss remaining challenges and propose a task execution framework that can efficiently support tiny tasks.
UR - http://www.scopus.com/inward/record.url?scp=85084189644&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85084189644&partnerID=8YFLogxK
M3 - Paper
AN - SCOPUS:85084189644
T2 - 14th Workshop on Hot Topics in Operating Systems, HotOS 2013
Y2 - 13 May 2013 through 15 May 2013
ER -