Authors: Emanuele Artioli (Alpen-Adria Universität Klagenfurt, Austria), Farzad Tashtarian (Alpen-Adria Universität Klagenfurt, Austria), Christian Timmerer (Alpen-Adria Universität Klagenfurt, Austria)

Venue: ACM 35th Workshop on Network and Operating System Support for Digital Audio and Video (NOSSDAV’25)

Abstract: The primary challenge of video streaming is to balance high video quality with smooth playback. Traditional codecs are well tuned for this trade-off, yet their inability to use context means they must encode the entire video data and transmit it to the client.
This paper introduces ELVIS (\textbf{E}nd-to-end \textbf{L}earning-based \textbf{VI}deo \textbf{S}treaming Enhancement Pipeline), an end-to-end architecture that combines server-side encoding optimizations with client-side generative in-painting to remove and reconstruct redundant video data. Its modular design allows ELVIS to integrate different codecs, in-painting models, and quality metrics, making it adaptable to future innovations.
Our results show that current technologies achieve improvements of up to 11 VMAF points over baseline benchmarks, though challenges remain for real-time applications due to computational demands. ELVIS represents a foundational step toward incorporating generative AI into video streaming pipelines, enabling higher quality experiences without increased bandwidth requirements.
By leveraging generative AI, we aim to develop a client-side tool, to incorporate in a dedicated video streaming player, that combines the accessibility of multilingual dubbing with the authenticity of the original speaker’s performance, effectively allowing a single actor to deliver their voice in any language. To the best of our knowledge, no current streaming system can capture the speaker’s unique voice or emotional tone.

 

We are glad that the paper was accepted for publication in Future Generation Computer Systems. This journal publishes cutting-edge research on high-performance computing, distributed systems, and advanced computing technologies for future computing environments.

Authors: Juan José Escobar, Pablo Sánchez-Cuevas, Beatriz Prieto, Rukiye Savran Kızıltepe, Fernando Díaz-del-Río, Dragi Kimovski

Abstract: Time and energy efficiency is a highly relevant objective in high-performance computing systems, with high costs for executing the tasks. Among these tasks, evolutionary algorithms are of consideration due to their inherent parallel scalability and usually costly fitness evaluation functions. In this respect, several scheduling strategies for workload balancing in heterogeneous systems have been proposed in the literature, with runtime and energy consumption reduction as their goals. Our hypothesis is that a dynamic workload distribution can be fitted with greater precision using metaheuristics, such as genetic algorithms, instead of linear regression. Therefore, this paper proposes a new mathematical model to predict the energy-time behaviour of applications based on multi-population genetic algorithms, which dynamically distributes the evaluation of individuals among the CPU-GPU devices of heterogeneous clusters. An accurate predictor would save time and energy by selecting the best resource set before running such applications. The estimation of the workload distributed to each device has been carried out by simulation, while the model parameters have been fitted in a two-phase run using another genetic algorithm and the experimental energy-time values of the target application as input. When the new model is analysed and compared with another based on linear regression, the one proposed in this work significantly improves the baseline approach, showing normalised prediction errors of 0.081 for runtime and 0.091 for energy consumption, compared to 0.213 and 0.256 shown in the baseline approach.

We are glad that the paper was accepted for publication in SCSA Journal. The journal covers research in smart computing systems and applications, with a focus on next-generation networking, cloud, and edge computing solutions.

Authors: Stojan Kitanov, Dragi Kimovski, Fisnik Doko, Kurt Horvath, Shpresa Tuda, Blerta Idrizi

Abstract: The rapid proliferation of IoT devices, coupled with the generated exponential growth of data, has necessitated the development of advanced network architectures. As a result, 5G mobile networks have already begun to face challenges such as network congestion, latency, and scalability limitations. Therefore, the need for a robust and future-proof solution becomes increasingly evident. In this direction, many research initiatives and industrial communities started to work on the development of 6G mobile networks. On the other hand, the emerging concept of Computing Continuum encompasses the seamless integration of edge, fog, and cloud computing resources to provide a unified and distributed computing environment, and it aims to enable real-time data processing, low-latency response, and intelligent decision-making at the network edge. The primary objective of this research paper is to address the shortcomings of existing network infrastructures and to overcome these shortcomings by integrating advanced AI capabilities in 6G mobile networks with the Computing Continuum. Moreover, it would be proposed a Computing Continuum Middleware for Artificial Intelligence over 6G networks, which would offer high-level and well-defined (“standardized”) interfaces which would create an automated, sustainable loop for managing IoT applications utilizing AI approaches over 6G networks.

We are glad that the paper was accepted for publication at ICFEC 2025. ICFEC focuses on innovations in cloud and edge computing, bringing together researchers and practitioners to discuss emerging challenges and solutions.

Title: ADApt: Edge Device Anomaly Detection and Microservice Replica Prediction

Authors: Narges Mehran, Nikolay Nikolov, Radu Prodan, Dumitru Roman, Dragi Kimovski, Frank Pallas, Peter Dorfinger

Venue: 9th IEEE International Conference on Fog and Edge Computing 2025, in conjunction with CCGrid 2025, 19-22 May, 2025 – Tromso, Norway

Abstract: The recent shift towards increasing user microservices in the Edge computing infrastructure brings new orchestration challenges, such as detecting overutilized resources and scaling out overloaded microservices in response to augmenting requests. In this work, we present the ADApt using the monitoring data related to Edge computing resources, detecting the utilization-based anomalies of resources (e.g., CORE or MEM), investigating the scalability in microservices, and adapting the application executions. To reduce the bottleneck while using computing resources, we first explore monitored devices executing microservices with various requirements, detecting overutilization-based processing events, and scoring them. Thereafter, based on the memory requirements, ADApt predicts the processing requirements of the microservices and estimates the number of replicas running on the overutilized devices. The prediction results show that the gradient boosting regression-based replica prediction reduces the MAE, MAPE, and RMSE compared to other models. Moreover, ADApt is able to estimate the number of replicas for each microservice close to the actual data without any prediction and to reduce the utilization of the device.

Authors: Reza Farahani, Radu Prodan

Venue: Workshop on AI for Sustainable Energy Systems and Green AI, Brdo Estate, Slovenia, 11-12 March 2025

 

 

Hadi

Authors: Fan Chen (Southwest Jiaotong University, China),  Lingfeng Qu (Guangzhou University, China), Hadi Amirpour (Alpen-Adria-Universität Klagenfurt), Christian Timmerer (Alpen-Adria-Universität Klagenfurt), and Hongjie He (Southwest Jiaotong University, China)

Journal: ACM Transactions on Multimedia Computing Communications and Applications (ACM TOMM)

Abstract: Reversible data hiding in encrypted images (RDH-EI) has gained widespread attention due to its potential applications in secure cloud storage. However, the security challenges of RDH-EI in cloud storage scenarios remain largely unexplored.} In this paper, we present a counterfeiting attack on RDH-EI schemes that utilize block-permutation and Co-XOR (BPCX) encryption. We demonstrate that ciphertext images generated by BPCX-based RDH-EI are easily tampered with to produce a counterfeit decrypted image with different contents imperceptible to the human eye. This vulnerability is mainly because the block permutation key information of BPCX is susceptible to known-plaintext attacks (KPAs). Taking ciphertext images in telemedicine scenarios as an example, we describe two potential counterfeiting attacks, namely fixed-area and optimal-area attacks. We show that the quality of forged decrypted images depends on the accuracy of the estimated block-permutation key under KPA conditions. To improve the invisibility of counterfeit decrypted images, we analyze the limitations of existing KPA methods against BPCX encryption for 2×2 block sizes and propose a novel diagonal inversion rule specifically designed for image blocks. This rule further enhances the accuracy of the estimated block-permutation key. The experiments show that, compared to existing KPA methods, the accuracy of the estimated block-permutation key in the UCID dataset increases by an average of 11.5%. In the counterfeiting attack experiments on Camera’s encrypted image, we successfully tampered with over 80% of the pixels in the target area under the fixed-region attack. Additionally, we achieved a tampering success rate exceeding 90% in the optimal-region attack.

 

Hadi

Tutorial title: Video Coding Advancements in HTTP Adaptive Streaming

Venue: IEEE International Conference on Multimedia & Expo (ICME) 2025 (https://2025.ieeeicme.org/tutorials/)

We are happy to announce that our tutorial “Video Coding Advancements in HTTP Adaptive Streaming” (by Hadi Amirpourazarian and Christian Timmerer) has been accepted for IEEE ICME 2025, which will take place in Nantes, France, June 30 – July 4, 2025.

Description: This tutorial provides a comprehensive exploration of the HTTP Adaptive Streaming (HAS) pipeline, covering advancements from content provisioning to content consumption. We begin by tracing the history of video streaming and the evolution of video coding technologies. Attendees will gain insights into the timeline of significant developments, from early proprietary solutions to modern adaptive streaming standards like HAS. A comparative analysis of video codecs is presented, highlighting milestones such as H.264, HEVC, and the latest standard, Versatile Video Coding (VVC), emphasizing their efficiency, adoption, and impact on streaming technologies. Additionally, new trends in video coding, including AI-based coding solutions, will be covered, showcasing their potential to transform video compression and streaming workflows.

Building on this foundation, we explore per-title encoding techniques, which dynamically tailor bitrate ladders to the specific characteristics of video content. These methods account for factors such as spatial resolution, frame rate, device compatibility, and energy efficiency, optimizing both Quality of Experience (QoE) and environmental sustainability. Next, we highlight cutting-edge advancements in live streaming, including novel approaches to optimizing bitrate ladders without introducing latency. Fast multi-rate encoding methods are also presented, showcasing how they significantly reduce encoding times and computational costs, effectively addressing scalability challenges for streaming providers.

The tutorial further delves into edge computing capabilities for video transcoding, emphasizing how edge-based architectures can streamline the processing and delivery of streaming content. These approaches reduce latency and enable efficient resource utilization, particularly in live and interactive streaming scenarios.

Finally, we discuss the QoE parameters that influence both streaming and coding pipelines, providing a holistic view of how QoE considerations guide decisions in codec selection, bitrate optimization, and delivery strategies. By combining historical context, theoretical foundations, and practical insights, this tutorial equips attendees with the knowledge to navigate and address the evolving challenges in video streaming applications.

Efficient Location-Based Service Discovery for IoT and Edge Computing in the 6G Era

Authors: Kurt Horvath, Dragi Kimovski

Conference: 2025 10th International Conference on Information and Network Technologies (ICINT 2025)

Abstract: Efficient service discovery is a cornerstone of the rapidly expanding Internet of Things (IoT) and edge computing ecosystems, where low latency and localized service provisioning are critical. This paper proposes a novel location-based DNS (Domain Name System) method that leverages Location Resource Records (LOC RRs) to enhance service discovery. By embedding geographic data in DNS responses, the system dynamically allocates services to edge nodes based on user proximity, ensuring reduced latency and improved Quality of Service (QoS). Comprehensive evaluations demonstrate minimal computational overhead, with processing times below 1 ms, making the approach highly suitable for latency-sensitive applications. Furthermore, the proposed methodology aligns with emerging 6G standards, which promise sub-millisecond latency and robust connectivity. Future research will focus on real-world deployment, validating the approach in dynamic IoT environments. This work establishes a scalable, efficient, and practical framework for location aware service discovery, providing a strong foundation for next generation IoT and edge-computing solutions.

 

Enhancing Traffic Safety with AI and 6G: Latency Requirements and Real-Time Threat Detection

Authors: Kurt Horvath, Dragi Kimovski, Stojan Kitanov, Radu Prodan

Conference: 2025 10th International Conference on Information and Network Technologies (ICINT 2025)

Abstract: The rapid digitalization of urban infrastructure opens the path to smart cities, where IoT-enabled infrastructure enhances public safety and efficiency. This paper presents a 6G and AI-enabled framework for traffic safety enhancement, focusing on real-time detection and classification of emergency vehicles and leveraging 6G as the latest global communication standard. The system integrates sensor data acquisition, convolutional neural network-based threat detection, and user alert dissemination through various software modules of the use case. We define the latency requirements for such a system, segmenting the end-toend latency into computational and networking components. Our empirical evaluation demonstrates the impact of vehicle speed and user trajectory on system reliability. The results provide insights for network operators and smart city service providers, emphasizing the critical role of low-latency communication and how networks can enable relevant services for traffic safety.

Tutorial title: Serverless Orchestration on the Edge-Cloud Continuum: Challenges and Solutions

Venue: 16th ACM/SPEC International Conference on Performance Engineering (ICPE) (https://icpe2025.spec.org/)

We are happy to announce that our tutorial “Serverless Orchestration on the Edge-Cloud Continuum: Challenges and Solutions” (by Reza Farahani and Radu Prodan) has been accepted for ACM/SPEC ICPE 2025, which will take place in Torento, Canada, in May 2025.

 

Authors: Sahar Nasirihaghighi, Negin Ghamsarian, Raphael Sznitman, Klaus Schoeffmann

Event: International Symposium on Biomedical Imaging (ISBI), April 14-17, 2025

Abstract: Accurate surgical phase recognition is crucial for advancing computer-assisted interventions, yet the scarcity of labeled data hinders training reliable deep learning models. Semi-supervised learning (SSL), particularly with pseudo-labeling, shows promise over fully supervised methods but often lacks reliable pseudo-label assessment mechanisms. To address this gap, we propose a novel SSL framework, Dual Invariance Self-Training (DIST), that incorporates both Temporal and Transformation Invariance to enhance surgical phase recognition. Our two-step self-training process dynamically selects reliable pseudo-labels, ensuring robust pseudo-supervision. Our approach mitigates the risk of noisy pseudo-labels, steering decision boundaries toward true data distribution and improving generalization to unseen data. Evaluations on Cataract and Cholec80 datasets show our method outperforms state-of-the-art SSL approaches, consistently surpassing both supervised and SSL baselines across various network architectures.