The last few years have seen a growing demand of distributed Cloud infrastructures able to process big data generated by geographically scattered sources. A key challenge of this environment is how to manage big data across multiple heterogeneous datacenters interconnected through imbalanced network links. We designed a Hierarchical Hadoop Framework (H2F) where a top-level business logic smartly schedules bottom-level computing tasks capable of exploiting the potential of the MapReduce within each datacenter. In this work we discuss on the opportunity of fragmenting the big data into small pieces so that better workload configurations may be devised for the bottom-level tasks. Several case study experiments were run on a testbed where a software prototype of the designed framework was deployed. The test results are reported and discussed in the last part of the paper.
|Titolo:||Fragmenting Big Data to Boost the Performance of MapReduce in Geographical Computing Contexts|
|Data di pubblicazione:||2017|
|Appare nelle tipologie:||4.1 Contributo in Atti di convegno|