Publication | Closed Access
Large-scale compute-intensive analysis via a combined in-situ and co-scheduling workflow approach
32
Citations
26
References
2015
Year
Unknown Venue
Cluster ComputingEngineeringComputer ArchitectureHigh Performance ComputingLarge-scale SimulationsGpu ComputingData ScienceSystems EngineeringModeling And SimulationParallel ComputingHigh-throughput ComputingMassively-parallel ComputingComputer EngineeringComputer ScienceCo-scheduling Workflow ApproachData-intensive ComputingWorkflow ExecutionProgram AnalysisLarge-scale Compute-intensive AnalysisCloud ComputingParallel ProgrammingFile SystemAnalysis RoutinesBig Data
Large-scale simulations can produce hundreds of terabytes to petabytes of data, complicating and limiting the efficiency of workflows. Traditionally, outputs are stored on the file system and analyzed in post-processing. With the rapidly increasing size and complexity of simulations, this approach faces an uncertain future. Trending techniques consist of performing the analysis in-situ, utilizing the same resources as the simulation, and/or off-loading subsets of the data to a compute-intensive analysis system. We introduce an analysis framework developed for HACC, a cosmological N-body code, that uses both in-situ and co-scheduling approaches for handling petabyte-scale outputs. We compare different analysis set-ups ranging from purely off-line, to purely in-situ to in-situ/co-scheduling. The analysis routines are implemented using the PISTON/VTK-m framework, allowing a single implementation of an algorithm that simultaneously targets a variety of GPU, multi-core, and many-core architectures.
| Year | Citations | |
|---|---|---|
Page 1
Page 1