Load-balancing algorithms in cloud computing: A survey
Cloud computing is a modern paradigm to provide services through the Internet. Load
balancing is a key aspect of cloud computing and avoids the situation in which some nodes …
balancing is a key aspect of cloud computing and avoids the situation in which some nodes …
A comprehensive view of Hadoop research—A systematic literature review
Context: In recent years, the valuable knowledge that can be retrieved from petabyte scale
datasets–known as Big Data–led to the development of solutions to process information …
datasets–known as Big Data–led to the development of solutions to process information …
A survey of large-scale analytical query processing in MapReduce
Enterprises today acquire vast volumes of data from different sources and leverage this
information by means of data analysis to support effective decision-making and provide new …
information by means of data analysis to support effective decision-making and provide new …
An experimental survey on big data frameworks
Recently, increasingly large amounts of data are generated from a variety of sources.
Existing data processing technologies are not suitable to cope with the huge amounts of …
Existing data processing technologies are not suitable to cope with the huge amounts of …
Silicon debug: scan chains alone are not enough
GJ Van Rootselaar, B Vermeulen - … Test Conference 1999 …, 1999 - ieeexplore.ieee.org
For today's multi-million transistor designs, existing design verification techniques cannot
guarantee that first silicon is designed error free. Therefore, techniques are necessary to …
guarantee that first silicon is designed error free. Therefore, techniques are necessary to …
Hadoop's adolescence: An analysis of Hadoop usage in scientific workloads
We analyze Hadoop workloads from three di? erent research clusters from a user-centric
perspective. The goal is to better understand data scientists' use of the system and how well …
perspective. The goal is to better understand data scientists' use of the system and how well …
Sailfish: A framework for large scale data processing
In this paper, we present Sailfish, a new Map-Reduce framework for large scale data
processing. The Sailfish design is centered around aggregating intermediate data …
processing. The Sailfish design is centered around aggregating intermediate data …
SHadoop: Improving MapReduce performance by optimizing job execution mechanism in Hadoop clusters
R Gu, X Yang, J Yan, Y Sun, B Wang, C Yuan… - Journal of parallel and …, 2014 - Elsevier
As a widely-used parallel computing framework for big data processing today, the Hadoop
MapReduce framework puts more emphasis on high-throughput of data than on low-latency …
MapReduce framework puts more emphasis on high-throughput of data than on low-latency …
Balancing reducer skew in MapReduce workloads using progressive sampling
The elapsed time of a parallel job depends on the completion time of its longest running
constituent. We present a static load balancing algorithm that distributes work evenly across …
constituent. We present a static load balancing algorithm that distributes work evenly across …
FiDoop-DP: data partitioning in frequent itemset mining on hadoop clusters
Traditional parallel algorithms for mining frequent itemsets aim to balance load by equally
partitioning data among a group of computing nodes. We start this study by discovering a …
partitioning data among a group of computing nodes. We start this study by discovering a …