The concept of scientific workflow makes it possible to link and control different tasks to carry out a complex treatment. The complicated workflow is generated by scientific distributed …
Task scheduling is a crucial key component for the efficient execution of data-intensive applications on distributed environments, by which many machines must be coordinated to …
The growing demands for data processing by new data-intensive applications are putting pressure on the performance and capacity of HPC storage systems. The advancement in …
As data intensive scientific computing systems become more widespread, there is a necessity of simplifying the development, deployment, and execution of complex data …
Computer applications are growing in terms of data management requirements. In both scientific and engineering domains, high-performance computing clusters tend to …
The ever-increasing power of supercomputer systems is both driving and enabling the emergence of new problem-solving methods that require the efficient execution of many …
Abstract The Data Mining Cloud Framework (DMCF) is an environment for designing and executing data analysis workflows in cloud platforms. Currently, DMCF relies on the default …
This paper explores novel techniques for improving the performance of many-task workflows based on the Swift scripting language. We propose novel programmer options for automated …
A Bandyopadhyay - 2011 26th IEEE/ACM International …, 2011 - ieeexplore.ieee.org
Spectrum based fault localization techniques such as Tarantula and Ochiai calculate the suspiciousness score of a program statement using the number of failing and passing test …