Distributed and Parallel Systems

The Symposium “The Data Science and Artificial Intelligence (DSAI) carnival” took place on the 19th of October 2023 at the Wageningen University & Research Campus and was organized in collaboration with the Wageningen Data Competence Center (WDCC).

This symposium provided an in-depth examination of cutting-edge themes from areas such as the Web, Semantic Web, linked data and knowledge graphs, LLMs, MLOps, cloud computing, data infrastructures and data space, FAIR data management, and related developments.

Leading experts shared the latest research and applications in these areas, fostering collaboration and offering insights into emerging trends.

The event concluded with the inaugural lecture of Prof. Dr. Anna Fensel.

 

Delighted to host last #PlenaryMeeting of #Datacloud 2020 project. Final adjustments for tool integration and business cases deployments ahead.

Authors: Gregor Molan, Gregor Dolinar, Jovan Bojkovski, Radu Prodan, Andrea Borghesi, Martin Molan

Journal: IEEE Access

Purpose: The gap between software development requirements and the available resources of software developers continues to widen. This requires changes in the development and organization of software development.

Objectives: Presented is a model introducing a quantitative software development management methodology that estimates the relative importance and risk of functionality retention or abundance, which determines the final value of the software product.

Method: The final value of the software product is interpreted as a function of the requirements and functionalities, represented as a computational graph (called a software product graph). The software product graph allows the relative importance of functionalities to be estimated by calculating the corresponding partial derivatives of the value function. The risk of not implementing the functionality is estimated by reducing the final value of a product.

Validation: This model has been applied to two EU projects: CareHD and vINCI. In vINCI, the functionalities with the most significant added value to the application were developed based on the implemented model and those that brought the least value were abandoned. Optimization was not implemented in the CareHD project and proceeded as initially designed. Consequently, only 71% of the CareHD’s potential value has been realized.

Conclusions: Presented model enables rational management and organization of software product development with real-time quantitative evaluation of functionalities impacts, assessment of the risks of omitting them without a significant impact. A quantitative evaluation of the impacts and risks of retention or abundance is possible based on the proposed algorithm, which is the core of the model. This model is a tool for rational organization and development of software products.

The GraphMassivizer plenary meeting is underway in #Walldorf, hosted by our valued partner metaphacts GmbH

We’re thrilled to gather for three days of collaboration, sharing project results, and planning upcoming activities.

Radu Prodan, Laurentiu Vasiliu, Irina Schmidt, Roberta Turra, Peter Haase, Richard Lloyd Stevens, Alexandru Iosup, Ana-Lucia Varbanescu, Nuria De Lama, Reza Farahani.

Title: Designing A Sustainable Serverless Graph Processing Tool on the Computing Continuum

Abstract: Graph processing has become increasingly popular and essential for solving complex problems in various domains, like social networks. However, processing graphs at a massive scale poses critical challenges, such as inefficient resource and energy utilization. To bridge such challenges, the Graph-Massivizer project, funded by the Horizon Europe research and innovation program, conducts research and develops a high-performance, scalable, and sustainable platform for information processing and reasoning based on the massive graph (MG) representation of extreme data. This paper presents an initial architectural design for the Choreographer, one of the five Graph-Massivizer tools. We explain Choreographer’s components and their collaboration with other Graph-Massivizer tools. We demonstrate how Choreographer can adopt the emerging serverless computing paradigm to process Basic Graph Operations (BGOs) as serverless functions across the computing continuum efficiently. Moreover, we present an early vision of our federated Function-as-a-Service (FaaS) testbed, which will be used to conduct experiments and assess Choreographer performance.

 

A DEEP DIVE INTO VIDEO STREAMING AND GRAPH PROCESSING USE CASES

More information

On 22.08.2023, Reza Farahani successfully defended his doctoral studies with the thesis on the title: “Network-Assisted Delivery of Adaptive Video Streaming Services through CDN, SDN, and MEC” under the supervision of Univ.-Prof. DI Dr. Hermann Hellwagner and Univ.-Prof. DI Dr. Christian Timmerer at ITEC. His defense was chaired by Assoc. Prof. DI Dr. Klaus Schöffmann and examined by Prof. Dr. Tobias Hoßfeld (University of Würzburg, Germany) and Prof. Dr. Filip De Turck (Ghent University, Belgium).
During his doctoral study, he contributed to ATHENA and Graph Massivizer projects.
Reza will continue as a Postdoctoral researcher at ITEC in the Graph Massivizer project.

The abstract of his disseration is as follows:

Multimedia applications, mainly video streaming services, are currently the dominant source of network load worldwide. In recent VoD and live video streaming services, traditional streaming delivery techniques have been replaced by adaptive solutions based on the HTTP protocol. Current trends toward high-resolution and low-latency VoD and live video streaming pose new challenges to E2E bandwidth demand and have stringent delay requirements. To do this, video providers rely on CDNs to ensure that they provide scalable video streaming services. To support future streaming scenarios involving millions of users, it is necessary to increase the CDNs’ efficiency. It is agreed that these requirements may be satisfied by adopting emerging networking techniques to present Network Assisted Video Streaming (NAVS) methods. Motivated by this, this thesis goes one step beyond traditional pure client-based HAS algorithms by incorporating (an) in-network component(s) with a broader view of the network to present completely transparent NAVS solutions for HAS clients. Our first contribution concentrates on leveraging the capabilities of the SDN, NFV, and MEC paradigms to introduce ES-HAS and CSDN as edge- and SDN-assisted frameworks. ES-HAS and CSDN introduce VNFs named VRP servers at the edge of an SDN-enabled network to collect HAS clients’ requests and retrieve networking information. The SDN controller in these systems manages a single domain network. VRP servers perform optimization models as server/segment selection policies to serve clients’ requests with the shortest fetching time by selecting the most appropriate cache server/video segment quality or by reconstructing the requested quality through transcoding at the edge. Deployment of ES-HAS and CSDN on the cloud-based testbeds and estimation of users’ QoE using objective metrics demonstrates how clients’ requests can be served with higher QoE by 40% and lower bandwidth usage by 63% compared to state-of-the-art approaches. Our second contribution designs an architecture that simultaneously supports various types of video streaming (live and VoD), considering their versatile QoE and latency requirements. To this end, the SDN, NFV, and MEC paradigms are leveraged, and three VNFs, i.e., VPF, VCF, and VTF, are designed. We build a series of these function chains through the SFC paradigm, utilize all CDN and edge server resources, and present SARENA, an SFC-enabled architecture for adaptive video streaming applications. We equip SARENA’s SDN controller with a lightweight request scheduler and edge configurator to make it deployable in practical environments and to dynamically scale edge servers based on service requirements, respectively. Experimental results show that SARENA outperforms baseline schemes in terms of higher users’ QoE figures by 39.6%, lower E2E latency by 29.3%, and lower backhaul traffic usage by 30% for live and VoD services. Our third contribution aims to use the idle resources of edge servers and employ the capabilities of the SDN controller to establish a collaboration between edge servers in addition to collaboration between edge servers and the SDN controller. We introduce two collaborative edge-assisted frameworks named LEADER and ARARAT. LEADER utilizes sets of actions, presented in an Action Tree, formulates the problem as a central optimization model to enhance the HAS clients’ serving time, subject to the network’s and edge servers’ resource constraints, and proposes a lightweight heuristic algorithm to solve the model. ARARAT extends LEADER’s Action Tree, considers network cost in the optimization, devises multiple heuristic algorithms, and runs extensive scenarios. Evaluation results show that LEADER and ARARAT improve users’ QoE by 22%, decrease the streaming cost by 47%, and enhance network utilization by 13%, as compared to others. Our final contribution focuses on incorporating P2P networks and CDNs, utilizing NFV and edge computing techniques, and then presenting RICHTER and ALIVE as hybrid P2P-CDN frameworks. RICHTER and ALIVE particularly use HAS clients’ potential idle computational resources besides their available bandwidth to provide distributed video processing services, e.g., video transcoding and video super-resolution. Both frameworks introduce multi-layer architectures and design Action Trees that consider all feasible resources for serving clients’ requests with acceptable latency and quality. Moreover, RICHTER proposes an online learning method and ALIVE utilizes a lightweight algorithm distributed over in-network virtualized components, which are designed to play decision-maker roles in large-scale practical scenarios. Results show that RICHTER and ALIVE improve the users’ QoE by 22%, decrease cost incurred for the streaming service provider by 34%, shorten clients’ serving latency by 39%, enhance edge server energy consumption by 31%, and reduce backhaul bandwidth usage by 24% compared to the others.

An der Universität Klagenfurt wird aktuell daran geforscht, wie große Datenmengen energieeffizienter verarbeitet werden können. Die digitale Übertragung von Informationen verbrauche Energie. Wissenschafter aus zwölf Institutionen forschen an der effizienteren Verarbeitung dieser sogenannten “massive graphs”, wie es in einer Aussendung des Projektteams am Mittwoch hieß. Ziel sei unter anderem ein Energielabel für Software-Codes einzuführen.

++ THEMENBILD ++ Illustration zum Thema Hacker, Hackerangriff und ComputerkriminalitŠt, fotografiert am 22. November 2022, in Wien. Angriffe auf IT-Infrastrukturen durch Cybercrime bereiten zunehmend Sorge.

“Das Sparpotenzial in der Verarbeitung von Daten wird noch zu wenig gesehen. Wir wollen es sichtbar machen und Lösungen anbieten”, sagte Projektleiter Radu Prodan. Im Green Supercomputing ginge es darum, die Rechenleistung effizienter zu organisieren, sodass in Summe weniger Energie verbraucht wird. Die Forscher arbeiten seit fast einem Jahr an dem Projekt “Extreme and Sustainable Graph Processing for Urgent Societal Challenges in Europe” und konnte bereits erste Ergebnisse vorweisen, die bisher auf drei Veranstaltungen in Portugal, Rumänien und in den USA präsentiert wurden.


Title: ARTICONF Decentralized Social Media Platform for Democratic Crowd Journalism

Authors: Ines Rito Lima, Vasco Filipe, Claudia Marinho, Alexandre Ulisses, Antorweep Chakravorty, Atanas Hristov, Nishant Saurabh, Zhiming Zhao, Ruyue Xin, Radu Prodan

Social Network Analysis and Mining https://www.springer.com/journal/13278

Abstract: Media production and consumption behaviors are changing in response to new technologies and demands, giving birth to a new generation of social applications. Among them, crowd journalism represents a novel way of constructing democratic and trustworthy news relying on ordinary citizens arriving at breaking news locations and capturing relevant videos using their smartphones. The ARTICONF  project proposes a trustworthy, resilient, and globally sustainable toolset for developing decentralized applications (DApps) to address this need. Its goal is to overcome the privacy, trust, and autonomy-related concerns associated with proprietary social media platforms overflowed by fake news.

Leveraging the ARTICONF tools, we introduce a new DApp for crowd journalism called MOGPlay. MOGPlay collects and manages audio-visual content generated by citizens and provides a secure blockchain platform that rewards all stakeholders involved in professional news production.

Besides live streaming, MOGPlay offers a marketplace for audio-visual content trading among citizens and free journalists with an internal token ecosystem. We discuss the functionality and implementation of the MOGPlay DApp and illustrate four pilot crowd journalism live scenarios that validate the prototype.