The effectiveness and scalability of MapReduce-based implementations of complex data-intensive tasks depend on an even redistribution of data between map and reduce tasks. In the...
Background: Bioinformatic analyses typically proceed as chains of data-processing tasks. A pipeline, or ‘workflow’, is a well-defined protocol, with a specific structure defin...
A key challenge faced by large-scale, distributed applications in Grid environments is efficient, seamless data management. In particular, for applications that can benefit from a...
Jithendar Paladugula, Ming Zhao 0002, Renato J. O....
Background: With advances in high-throughput genomics and proteomics, it is challenging for biologists to deal with large data files and to map their data to annotations in public...
In designing dynamic situations such as cyberworlds, we the Incrementally Modular Abstraction Hierarchy (IMAH) to be an appropriate mathematical background to model dynamically ch...