The first review of the ASPIDE project took place on 25.02.2020 in the premises of the European Commission in Luxemburg. During the project review, a live demo of the platform for supporting extreme scale applications was presented and future research and developing activities were discussed with the reviewers.

Aspide-Review-2020

Aspide Review 2020

ARTICONF: EU first review

ARTICONF: EU first review

Bitmovin, a world leader in online video technology, is teaming up with the University of Klagenfurt, Institute of Information Technology (ITEC) and the Austrian Federal Ministry of Digital and Economic Affairs (BMDW) in a multi-million Euro research project to uncover techniques that will enhance the video streaming experiences of the future. The joint project establishes a dedicated research team to investigate potential new tools and methodologies for encoding, transport and playback of live and on-demand video using the HTTP Adaptive Streaming protocol that is widely used by online video and TV providers. The resulting findings will help empower the creation of next-generation solutions for higher quality video experiences at lower latency, while also potentially reducing storage and distribution costs.

Continue reading …

The ITEC team participated in the HiPeac 2020 International Workshop on Exascale programing models for extreme data with a presentation with title “Monitoring data collection and mining for Exascale systems”. The ITEC team also attended the collocated ASPIDE meeting and actively participated in the decision of the next research activities in the project.

Dragi Kimovski

Title of the talk: Mobility-Aware Scheduling of Extreme Data Workflows across the Computing Continuum

Abstract: The appearance of the Fog/Edge computing paradigm, as an emanation of the computing continuum closer to the edge of the network, unravels important opportunities for execution of complex business and scientific workflows near the data sources. The main characteristics of these workflows are (i) their distributed nature, (ii) the vast amount of data (in the order of petabytes) they generate and (iii) the strict latency requirements. Current workflow management approaches rely exclusively on the Cloud Data Centers, which due to their geographical distance in relation to the data sources, could negatively influence the latency and cause violation of workflow requirements. It is therefore essential to research novel concepts for partial offloading of complex workflows closer to where the data is generated, thus reducing the communication latency and the need for frequent data transfers.

In this talk we will explore the  potential  of  the computing continuum  for  scheduling and partial offloading  of  complex  workflows  with  strict  response time requirements and expose the resource provisioning challenges related to the heterogeneity and mobility of the Fog/Edge environment. Consequently, we will discuss a novel mobility-aware Pareto-based approach for task offloading across the continuum, which considers three optimization objectives, namely response time, reliability, and financial cost. Besides, the approach introduces a Markov model to perform a single-step predictive analysis on the mobility of the Fog/Edge devices, thus constraining the task offloading optimization problem to devices that do not frequently move (roam) within the computing continuum. As a conclusion to the talk, we will discuss the efficiency of the presented approach, based on both a simulated and a real-world testbed environment tailored for a set real-world biomedical, meteorological and astronomy workflows.

IWCoCo 2020 in Bologna

The manuscript “Simplified Workflow Simulation on Clouds based on Computation and Communication Noisiness” has been accepted for
publication in the IEEE Transactions on Parallel and Distributed Systems journal (TPDS) with an impact factor of 4.181.

Authors: Roland Mathá, Sasko Ristov, Thomas Fahringer, Radu Prodan.

Abstract: Many researchers rely on simulations to analyze and validate their researched methods on Cloud infrastructures. However, determining relevant simulation parameters and correctly instantiating them to match the real Cloud performance is a difficult and costly operation, as minor configuration changes can easily generate an unreliable inaccurate simulation result. Using legacy values experimentally determined by other researchers can reduce the configuration costs, but is still inaccurate as the underlying public Clouds and the number of active tenants are highly different and dynamic in time. To overcome these deficiencies, we propose a novel model that simulates the dynamic Cloud performance by introducing noise in the computation and communication tasks, determined by a small set of runtime execution data. Although the estimating method is apparently costly, a comprehensive sensitivity analysis shows that the configuration parameters determined for a certain simulation setup can be used for other simulations too, thereby reducing the tuning cost by up to 82.46%, while declining the simulation accuracy by only 1.98% in average. Extensive evaluation also shows that our novel model outperforms other state-of-the-art dynamic Cloud simulation models, leading up to 22% lower makespan inaccuracy.

Acknowledgments: This work has been supported by the ASPIDE Project funded by the European Union’s Horizon 2020 Research and Innovation Programme under grant agreement No 801091.

An der Universität Klagenfurt wird das CD Labor für Adaptive Streaming over HTTP and Emerging Networked Multimedia Services eingerichtet. Die Mission des Labors ist die Erforschung neuer Tools und Methoden für die Codierung, den Transport und die Wiedergabe von Live- und On-Demand-Videos mithilfe des HTTP-Adaptive-Streaming-Verfahrens. In Christian Doppler Labors wird anwendungsorientierte Grundlagenforschung auf hohem Niveau betrieben, hervorragende WissenschaftlerInnen kooperieren dazu mit innovativen Unternehmen. Für die Förderung dieser Zusammenarbeit gilt die Christian Doppler Forschungsgesellschaft international als Best-Practice-Beispiel. Christian Doppler Labors werden von der öffentlichen Hand und den beteiligten Unternehmen gemeinsam finanziert. Wichtigster öffentlicher Fördergeber ist das Bundesministerium für Digitalisierung und Wirtschaftsstandort (BMDW). Wir laden Sie herzlich zur feierlichen Eröffnung des CD Labors ein. Read more