dc.contributor.author | Nikolov, Nikolay | |
dc.contributor.author | Dessalk, Yared Dejene | |
dc.contributor.author | Khan, Akif Quddus | |
dc.contributor.author | Soylu, Ahmet | |
dc.contributor.author | Matskin, Mihhail | |
dc.contributor.author | Payberah, Amir | |
dc.contributor.author | Roman, Dumitru | |
dc.date.accessioned | 2022-03-07T12:45:30Z | |
dc.date.available | 2022-03-07T12:45:30Z | |
dc.date.created | 2021-11-04T11:18:06Z | |
dc.date.issued | 2021-11-26 | |
dc.identifier.citation | Internet of Things. 2021, 16 1-19. | en_US |
dc.identifier.issn | 2542-6605 | |
dc.identifier.uri | https://hdl.handle.net/11250/2983427 | |
dc.description.abstract | Big Data processing, especially with the increasing proliferation of Internet of Things (IoT) technologies and convergence of IoT, edge and cloud computing technologies, involves handling massive and complex data sets on heterogeneous resources and incorporating different tools, frameworks, and processes to help organizations make sense of their data collected from various sources. This set of operations, referred to as Big Data workflows, requires taking advantage of Cloud infrastructures’ elasticity for scalability. In this article, we present the design and prototype implementation of a Big Data workflow approach based on the use of software container technologies, message-oriented middleware (MOM), and a domain-specific language (DSL) to enable highly scalable workflow execution and abstract workflow definition. We demonstrate our system in a use case and a set of experiments that show the practical applicability of the proposed approach for the specification and scalable execution of Big Data workflows. Furthermore, we compare our proposed approach’s scalability with that of Argo Workflows – one of the most prominent tools in the area of Big Data workflows – and provide a qualitative evaluation of the proposed DSL and overall approach with respect to the existing literature. | en_US |
dc.description.sponsorship | This work was partly funded by the EC H2020 projects ‘‘DataCloud: Enabling The Big Data Pipeline Lifecycle on the Computing Continuum’’ (Grant nr. 101016835) and ‘‘COGNITWIN: Cognitive plants through proactive self-learning hybrid digital twins’’ (Grant nr. 870130), and the NFR project ‘‘BigDataMine’’ (Grant nr. 309691). | en_US |
dc.language.iso | eng | en_US |
dc.publisher | Elsevier | en_US |
dc.relation.ispartofseries | Internet of Things;Volume 16, December 2021, 100440 | |
dc.rights | Navngivelse 4.0 Internasjonal | * |
dc.rights.uri | http://creativecommons.org/licenses/by/4.0/deed.no | * |
dc.subject | Internet of Things | en_US |
dc.subject | Big Data | en_US |
dc.subject | Big data workflows | en_US |
dc.subject | Domain-specific languages | en_US |
dc.subject | Software containers | en_US |
dc.title | Conceptualization and scalable execution of big data workflows using domain-specific languages and software containers | en_US |
dc.type | Peer reviewed | en_US |
dc.type | Journal article | en_US |
dc.description.version | publishedVersion | en_US |
dc.rights.holder | © 2021 The Authors | en_US |
dc.source.articlenumber | 100440 | en_US |
cristin.ispublished | true | |
cristin.fulltext | postprint | |
cristin.fulltext | original | |
cristin.qualitycode | 1 | |
dc.identifier.doi | https://doi.org/10.1016/j.iot.2021.100440 | |
dc.identifier.cristin | 1951355 | |
dc.source.journal | Internet of Things | en_US |
dc.source.volume | 16 | en_US |
dc.source.pagenumber | 1-19 | en_US |
dc.relation.project | Norges forskningsråd: 309691 | en_US |
dc.relation.project | EC/H2020/870130 | en_US |
dc.relation.project | EC/H2020/101016835 | en_US |
dc.subject.nsi | VDP::Datateknologi: 551 | en_US |
dc.subject.nsi | VDP::Computer technology: 551 | en_US |
dc.subject.nsi | VDP::Datateknologi: 551 | en_US |
dc.subject.nsi | VDP::Computer technology: 551 | en_US |
dc.subject.nsi | VDP::Datateknologi: 551 | en_US |
dc.subject.nsi | VDP::Computer technology: 551 | en_US |