How can we optimally trade extra computing power to reduce the communication load in distributed computing? We answer this question by characterizing a fundamental tradeoff between computation and ...communication in distributed computing, i.e., the two are inversely proportional to each other. More specifically, a general distributed computing framework, motivated by commonly used structures like MapReduce, is considered, where the overall computation is decomposed into computing a set of "Map" and "Reduce" functions distributedly across multiple computing nodes. A coded scheme, named "coded distributed computing" (CDC), is proposed to demonstrate that increasing the computation load of the Map functions by a factor of r (i.e., evaluating each function at r carefully chosen nodes) can create novel coding opportunities that reduce the communication load by the same factor. An information-theoretic lower bound on the communication load is also provided, which matches the communication load achieved by the CDC scheme. As a result, the optimal computation-communication tradeoff in distributed computing is exactly characterized. Finally, the coding techniques of CDC is applied to the Hadoop TeraSort benchmark to develop a novel CodedTeraSort algorithm, which is empirically demonstrated to speed up the overall job execution by 1.97× -3.39×, for typical settings of interest.
Cloud computing uses the concepts of scheduling and load balancing to migrate tasks to underutilized VMs for effectively sharing the resources. The scheduling of the nonpreemptive tasks in the cloud ...computing environment is an irrecoverable restraint and hence it has to be assigned to the most appropriate VMs at the initial placement itself. Practically, the arrived jobs consist of multiple interdependent tasks and they may execute the independent tasks in multiple VMs or in the same VM’s multiple cores. Also, the jobs arrive during the run time of the server in varying random intervals under various load conditions. The participating heterogeneous resources are managed by allocating the tasks to appropriate resources by static or dynamic scheduling to make the cloud computing more efficient and thus it improves the user satisfaction. Objective of this work is to introduce and evaluate the proposed scheduling and load balancing algorithm by considering the capabilities of each virtual machine (VM), the task length of each requested job, and the interdependency of multiple tasks. Performance of the proposed algorithm is studied by comparing with the existing methods.
Motivated by the recent explosion of interest around blockchains, we examine whether they make a good fit for the Internet of Things (IoT) sector. Blockchains allow us to have a distributed ...peer-to-peer network where non-trusting members can interact with each other without a trusted intermediary, in a verifiable manner. We review how this mechanism works and also look into smart contracts-scripts that reside on the blockchain that allow for the automation of multi-step processes. We then move into the IoT domain, and describe how a blockchain-IoT combination: 1) facilitates the sharing of services and resources leading to the creation of a marketplace of services between devices and 2) allows us to automate in a cryptographically verifiable manner several existing, time-consuming workflows. We also point out certain issues that should be considered before the deployment of a blockchain network in an IoT setting: from transactional privacy to the expected value of the digitized assets traded on the network. Wherever applicable, we identify solutions and workarounds. Our conclusion is that the blockchain-IoT combination is powerful and can cause significant transformations across several industries, paving the way for new business models and novel, distributed applications.
We present a methodology for defining and optimizing a general force field for classical molecular simulations, and we describe its use to derive the Open Force Field 1.0.0 small-molecule force ...field, codenamed Parsley. Rather than using traditional atom typing, our approach is built on the SMIRKS-native Open Force Field (SMIRNOFF) parameter assignment formalism, which handles increases in the diversity and specificity of the force field definition without needlessly increasing the complexity of the specification. Parameters are optimized with the ForceBalance tool, based on reference quantum chemical data that include torsion potential energy profiles, optimized gas-phase structures, and vibrational frequencies. These quantum reference data are computed and are maintained with QCArchive, an open-source and freely available distributed computing and database software ecosystem. In this initial application of the method, we present essentially a full optimization of all valence parameters and report tests of the resulting force field against compounds and data types outside the training set. These tests show improvements in optimized geometries and conformational energetics and demonstrate that Parsley's accuracy for liquid properties is similar to that of other general force fields, as is accuracy on binding free energies. We find that this initial Parsley force field affords accuracy similar to that of other general force fields when used to calculate relative binding free energies spanning 199 protein-ligand systems. Additionally, the resulting infrastructure allows us to rapidly optimize an entirely new force field with minimal human intervention.
•An activity flow mapping approach was adopted.•To parse distributed versus local processing contributions to task-evoked activations•Activity flow prediction dissociates the load-dependent ...activations of DAN and ECN•Adaptive training selectively induces improved information integration for the ECN•A causal effect of activity flow prediction was revealed by training manipulation
Flexible cognitive functions, such as working memory (WM), usually require a balance between localized and distributed information processing. However, it is challenging to uncover how local and distributed processing specifically contributes to task-induced activity in a region. Although the recently proposed activity flow mapping approach revealed the relative contribution of distributed processing, few studies have explored the adaptive and plastic changes that underlie cognitive manipulation. In this study, we recruited 51 healthy volunteers (31 females) and investigated how the activity flow and brain activation of the frontoparietal systems was modulated by WM load and training. While the activation of both executive control network (ECN) and dorsal attention network (DAN) increased linearly with memory load at baseline, the relative contribution of distributed processing showed a linear response only in the DAN, which was prominently attributed to within-network activity flow. Importantly, adaptive training selectively induced an increase in the relative contribution of distributed processing in the ECN and also a linear response to memory load, which were predominantly due to between-network activity flow. Furthermore, we demonstrated a causal effect of activity flow prediction through training manipulation on connectivity and activity. In contrast with classic brain activation estimation, our findings suggest that the relative contribution of distributed processing revealed by activity flow prediction provides unique insights into neural processing of frontoparietal systems under the manipulation of cognitive load and training. This study offers a new methodological framework for exploring information integration versus segregation underlying cognitive processing.
The emergence of edge computing has witnessed a fast-growing volume of data on edge devices belonging to different stakeholders which, however, cannot be shared among them due to the lack of the ...trust. By exploiting blockchain's non-repudiation and non-tampering properties that enable trust, we develop a blockchain-based big data sharing framework to support various applications across resource-limited edges. In particular, we devise a number of novel resource-efficient techniques for the framework: (1) the PoC (Proof-of-Collaboration) based consensus mechanism with low computation complexity which is especially beneficial to the edge devices with low computation capacity, (2) the blockchain transaction filtering and offloading scheme that can significantly reduce the storage overhead, and (3) new types of blockchain transaction (i.e., Express Transaction) and block (i.e., Hollow Block) to enhance the communication efficiency. Extensive experiments are conducted and the results demonstrate the superior performance of our proposal.
•A comprehensive review of the state-of-the-art blockchain consensus algorithms.•An analytical framework to evaluate the pros and cons of consensus mechanisms.•The comparison criteria are weighted by ...the pairwise comparison method.•The existing open issues, challenges and directions to enlighten future research.
How to reach an agreement in a blockchain network is a complex and important task that is defined as a consensus problem and has wide applications in reality including distributed computing, load balancing, and transaction validation in blockchains. Over recent years, many studies have been done to cope with this problem. In this paper, a comparative and analytical review on the state-of-the-art blockchain consensus algorithms is presented to enlighten the strengths and constraints of each algorithm. Based on their inherent specifications, each algorithm has a different domain of applicability that yields to propose several performance criteria for the evaluation of these algorithms. To overview and provide a basis of comparison for further work in the field, a set of incommensurable and conflicting performance evaluation criteria is identified and weighted by the pairwise comparison method. These criteria are classified into four categories including algorithms’ throughput, the profitability of mining, degree of decentralization and consensus algorithms vulnerabilities and security issues. Based on the proposed framework, the pros and cons of consensus algorithms are systematically analyzed and compared in order to provide a deep understanding of the existing research challenges and clarify the future study directions.