Enabling Scalable Data Processing and Management through Standards-based Job Execution and the Global Federated File System

dc.contributorHáskóli Íslandsen_US
dc.contributorUniversity of Icelanden_US
dc.contributor.authorMemon, Shahbaz
dc.contributor.authorRiedel, Morris
dc.contributor.authorMemon, Riedel,
dc.contributor.authorKoeritz, Chris
dc.contributor.authorGrimshaw, Andrew
dc.contributor.authorNeukirchen, Helmut
dc.contributor.departmentIðnaðarverkfræði-, vélaverkfræði- og tölvunarfræðideild (HÍ)en_US
dc.contributor.departmentFaculty of Industrial Eng., Mechanical Eng. and Computer Science (UI)en_US
dc.contributor.schoolVerkfræði- og náttúruvísindasvið (HÍ)en_US
dc.contributor.schoolSchool of Engineering and Natural Sciences (UI)en_US
dc.date.accessioned2017-01-30T11:06:02Z
dc.date.available2017-01-30T11:06:02Z
dc.date.issued2016-05-01
dc.description.abstractEmerging challenges for scientific communities are to efficiently process big data obtained by experimentation and computational simulations. Supercomputing architectures are available to support scalable and high performant processing environment, but many of the existing algorithm implementations are still unable to cope with its architectural complexity. One approach is to have innovative technologies that effectively use these resources and also deal with geographically dispersed large datasets. Those technologies should be accessible in a way that data scientists who are running data intensive computations do not have to deal with technical intricacies of the underling execution system. Our work primarily focuses on providing data scientists with transparent access to these resources in order to easily analyze data. Impact of our work is given by describing how we enabled access to multiple high performance computing resources through an open standards-based middleware that takes advantage of a unified data management provided by the the Global Federated File System. Our architectural design and its associated implementation is validated by a usecase that requires massivley parallel DBSCAN outlier detection on a 3D point clouds dataset.en_US
dc.description.versionAccepteden_US
dc.format.extent115-128en_US
dc.identifier.citationShahbaz Memon, Morris Riedel, Shiraz Memon, Chris Koeritz, Andrew Grimshaw, Helmut Neukirchen. (2016). Enabling Scalable Data Processing and Management through Standards-based Job Execution and the Global Federated File System. Scalable Computing: Practice and Experience, 17(2). 115-128. DOI: http://dx.doi.org/10.1051/kmae/2011046en_US
dc.identifier.doi10.12694/scpe.v17i2.1160
dc.identifier.issn1895-1767
dc.identifier.journalScalable Computing: Practice and Experienceen_US
dc.identifier.urihttps://hdl.handle.net/20.500.11815/184
dc.language.isoenen_US
dc.publisherWest University of Timisoaraen_US
dc.relation.ispartofseriesScalable Computing: Practice and Experience;17(2)
dc.relation.urlhttp://www.scpe.org/index.php/scpe/article/view/1160en_US
dc.rightsinfo:eu-repo/semantics/openAccessen_US
dc.subjectStatistical data miningen_US
dc.subjectData processing,en_US
dc.subjectDistributed file systemen_US
dc.subjectGagnavinnslaen_US
dc.subjectGagnanámen_US
dc.subjectSkráning gagnaen_US
dc.titleEnabling Scalable Data Processing and Management through Standards-based Job Execution and the Global Federated File Systemen_US
dc.typeinfo:eu-repo/semantics/articleen_US
dcterms.licenseOpen Accessen_US

Skrár

Original bundle

Niðurstöður 1 - 1 af 1
Hleð...
Thumbnail Image
Nafn:
1160-1099-1-PB.pdf
Stærð:
644.45 KB
Snið:
Adobe Portable Document Format
Description:
Publisher´s version

Undirflokkur