Title: FSpot: Fast Multi-Objective Heuristic for Efficient Video Encoding Workloads over AWS EC2 Spot Instance Fleet

Authors: Anatoliy Zabrovskiy, Prateek Agrawal, Vladislav Kashansky, Roland Kersche, Christian Timmerer, and Radu Prodan

Abstract: HTTP Adaptive Streaming (HAS) of video content is becoming an undivided part of the Internet and accounts for most of today’s network traffic. Video compression technology plays a vital role in efficiently utilizing network channels, but encoding videos into multiple representations with selected encoding parameters is a significant challenge. However, video encoding is a computationally intensive and time-consuming operation that requires high-performance resources provided by on-premise infrastructures or public clouds. In turn, the public clouds, such as Amazon elastic compute cloud (EC2), provide hundreds of computing instances optimized for different purposes and clients’ budgets. Thus, there is a need for algorithms and methods for optimized computing instance selection for specific tasks such as video encoding and transcoding operations. Additionally, the encoding speed directly depends on the selected encoding parameters and the complexity characteristics of video content. In this paper, we first benchmarked the video encoding performance of Amazon EC2 spot instances using multiple x264 codec encoding parameters and video sequences of varying complexity. Then, we proposed a novel fast approach to optimize Amazon EC2 spot instances and minimize video encoding costs. Furthermore, we evaluated how the optimized selection of EC2 spot instances can affect the encoding cost. The results show that our approach, on average, can reduce the encoding costs by at least 15.8% and up to 47.8% when compared to a random selection of EC2 spot instances.

Keywords: EC2 Spot instance, Encoding time prediction; adaptive streaming; video transcoding; Clustering; HTTP adaptive streaming; MPEG-DASH; Cloud computing; optimization; Pareto front.

The Fast Multi-Resolution and Multi-Rate Encoding for HTTP Adaptive Streaming Using Machine Learning paper from ATHENA lab has won the Best New Streaming Innovation Award in the Streaming Media Readers’ Choice Awards 2021.

The journey that led to the publication of the FaRes-ML paper was quite an insightful one.

It all started with the question, “How to efficiently provide multi-rate representations over a wide range of resolutions for HTTP Adaptive Streaming?“. This led to the first publication, Fast Multi-Rate Encoding for Adaptive HTTP Streaming, in which we proposed a double-bound approach to speed up the multi-rate encoding. After analyzing the results, we saw room for improvement in parallel encoding performance, which led to the second publication Towards Optimal Multirate Encoding for HTTP Adaptive Streaming. The results were promising, but we believed we could improve the encoding performance by utilizing machine learning. That was the primary motivation behind our third paper, FaME-ML: Fast Multirate Encoding for HTTP Adaptive Streaming Using Machine Learning. In FaMe-ML, we have used convolutional neural networks (CNNs) to use the information from the reference representation better to encode other representations, resulting in significant improvement in the multi-rate encoding performance. Finally, we proposed FaRes-ML to extend our FaME-ML approach to include multi-resolution scenarios in Fast Multi-Resolution and Multi-Rate Encoding for HTTP Adaptive Streaming Using Machine Learning paper.

Here is the list of publications that led to FaRes-ML:

  1. Fast Multi-Rate Encoding for Adaptive HTTP Streaming. Published in DCC’20.
  2. Towards Optimal Multirate Encoding for HTTP Adaptive Streaming. Published in MMM’21.
  3. FaME-ML: Fast Multirate Encoding for HTTP Adaptive Streaming Using Machine Learning. Published in VCIP’20.
  4. Fast Multi-Resolution and Multi-Rate Encoding for HTTP Adaptive Streaming Using Machine Learning. Published in IEEE OJ-SP.

Taichung, Taiwan, The 1st IEEE International Workshop on Data-Driven Rate Control for Media Streaming (DDRC’21) Co-located with the IEEE International Conference on Multimedia Big Data (BigMM’21)

Conference Website: https://www.bigmm.org/ (November 15-17)

HTTP Adaptive Streaming (HAS) — Quo Vadis?
Speaker: Professor Christian Timmerer
Time: November 16, 2021 12:10 (UTC +1)

CAdViSE or how to find the Sweet Spots of ABR Systems
Speaker: Babak Taraghi, M.Sc.
Time: November 16, 2021 13:00 (UTC +1)

Online attendance is free, Visit here for more information.

Over the past decade, Carinthia has developed into a hotspot in drone research. The heart of the scientific operation is Europe’s largest drone hall in Klagenfurt.

The newspaper “Stuttgarter Nachrichten” reported on our drone work with its own article (link).

28th International Conference on Multimedia Modeling (MMM)

April 05-08, 2022 | Qui Nhon, Vietnam

Conference Website

Jesús Aguilar Armijo (Alpen-Adria-Universität Klagenfurt), Ekrem Çetinkaya (Christian Doppler Laboratory ATHENA, Alpen-Adria-Universität Klagenfurt), Christian Timmerer (Alpen-Adria-Universität Klagenfurt) and Hermann Hellwagner (Alpen-Adria-Universität Klagenfurt)

Abstract: As the video streaming traffic in mobile networks is increasing, improving the content delivery process becomes crucial, e.g., by utilizing edge computing support. At an edge node, we can deploy adaptive
bitrate (ABR) algorithms with a better understanding of network behavior and access to radio and player metrics. In this work, we present ECAS-ML, Edge Assisted Adaptation Scheme for HTTP Adaptive Streaming
with Machine Learning. ECAS-ML focuses on managing the tradeoff among bitrate, segment switches and stalls to achieve a higher quality of experience (QoE). For that purpose, we use machine learning techniques
to analyze radio throughput traces and predict the best parameters of our algorithm to achieve better performance. The results show that ECAS-ML outperforms other client-based and edge-based ABR algorithms.

Keywords: HTTP Adaptive Streaming, Edge Computing, Content Delivery, Network-assisted Video Streaming, Quality of Experience, Machine Learning.

28th International Conference on Multimedia Modeling (MMM)

April 05-08, 2022 | Qui Nhon, Vietnam

Conference Website

Ekrem Çetinkaya (Christian Doppler Laboratory ATHENA, Alpen-Adria-Universität Klagenfurt), Minh Nguyen (Christian Doppler Laboratory ATHENA, Alpen-Adria-Universität Klagenfurt), and Christian Timmerer (Christian Doppler Laboratory ATHENA, Alpen-Adria-Universität Klagenfurt)

Abstract: Deep neural network (DNN) based approaches have been intensively studied to improve video quality thanks to their fast advancement in recent years. These approaches are designed mainly for desktop devices due to their high computational cost. However, with the increasing performance of mobile devices in recent years, it became possible to execute DNN based approaches in mobile devices. Despite having the required computational power, utilizing DNNs to improve the video quality for mobile devices is still an active research area. In this paper, we propose an open-source mobile platform, namely MoViDNN, to evaluate DNN based video quality enhancement methods, such as super-resolution, denoising, and deblocking. Our proposed platform can be used to evaluate the DNN based approaches both objectively and subjectively. For objective evaluation, we report common metrics such as execution time, PSNR, and SSIM. For subjective evaluation, Mean Score Opinion (MOS) is reported. The proposed platform is available publicly at https://github.com/cd-athena/MoViDNN

Keywords: Super resolution, Deblocking, Deep Neural Networks, Mobile Devices

The 1st ACM CoNEXT Workshop on Design, Deployment, and Evaluation of Network-assisted  video Streaming (ViSNext 2021)

Conference Website

Alireza Erfanian (Alpen-Adria-Universität Klagenfurt), Hadi Amirpour (Alpen-Adria-Universität Klagenfurt), Farzad Tashtarian (Alpen-Adria-Universität Klagenfurt), Christian Timmerer (Alpen-Adria-Universität Klagenfurt), and Hermann Hellwagner (Alpen-Adria-Universität Klagenfurt)

Abstract: Live video streaming is widely embraced in video services, and its applications have attracted much attention in recent years. The increased number of users demanding high quality (e.g., 4K resolution) live videos increase the bandwidth utilization in the backhaul network. To decrease bandwidth utilization in HTTP Adaptive Streaming (HAS), in on-the-fly transcoding approaches, only the highest bitrate representation is delivered to the edge, and other representations are generated by transcoding at the edge. However, this approach is inefficient due to the high transcoding cost. In this paper, we propose a light-weight transcoding at the edge method for live applications, LwTE-Live, to decrease the band-width utilization and the overall live streaming cost. During the encoding processes at the origin server, the optimal encoding decisions are saved as metadata, and the metadata replaces the corresponding representation in the bitrate ladder. The significantly reduced size of the metadata compared to its corresponding representation decreases the bandwidth utilization. The extracted metadata is then utilized at the edge to decrease the transcoding time. We formulate the problem as a Mixed-Binary Linear Programming (MBLP) model to optimize the live streaming cost, including the bandwidth and computation costs. We compare the proposed model with state-of-the-art approaches and the experimental results show that our proposed method saves the cost and backhaul bandwidth utilization up to 34% and 45%, respectively.

Keywords: live video streaming, network function virtualization, NFV, light-weight transcoding, transcoding, edge computing

The 10th IFIP/IEEE International Conference on Performance Evaluation and Modeling in Wired and Wireless Networks (PEMWN)

Conference Website

Jesús Aguilar Armijo (Alpen-Adria-Universität Klagenfurt), Christian Timmerer (Alpen-Adria-Universität Klagenfurt) and Hermann Hellwagner (Alpen-Adria-Universität Klagenfurt)

Abstract: In this work, we present ANGELA, HTTP adaptive streaming (HAS) and Edge Computing Simulator. ANGELA was designed to test edge mechanisms that support HAS, as it offers: realistic radio layer simulation, different multimedia content configurations, access to radio and player metrics at the edge, and a wide variety of metrics to evaluate the video streaming session performance. The ANGELA architecture is flexible and can support adaptive bitrate (ABR) algorithms located at different points of the network. Moreover, we show the possibilities of Angela by evaluating different ABR algorithms.

Keywords: Network simulator, testbed, edge computing, HTTP Adaptive Streaming.

The ACM CoNEXT 2021 Workshop on the Evolution, Performance, and Interoperability of QUIC (EPIQ)

07 December 2021  | Munich, Germany (Online)

Workshop Website

Daniele Lorenzi (Department of Information Engineering, University of PaduaItaly), Minh Nguyen (Christian Doppler Laboratory ATHENA, Alpen-Adria-Universität Klagenfurt), Farzad Tashtarian (Christian Doppler Laboratory ATHENA, Alpen-Adria-Universität Klagenfurt), Simone Milani (Department of Information Engineering, University of PaduaItaly), Herman Hellwagner (Christian Doppler Laboratory ATHENA, Alpen-Adria-Universität Klagenfurt),  Christian Timmerer (Christian Doppler Laboratory ATHENA, Alpen-Adria-Universität Klagenfurt)

Abstract: HTTP Adaptive Streaming(HAS) has become a predominant technique for delivering videos in the Internet. Due to its adaptive behaviour according to changing network conditions it may result in video quality variations that negatively impacts the Quality of Experience (QoE) of the user. In this paper, we propose Days of Future Past, an optimization-based Adaptive Bitrate (ABR) algorithm over HTTP/3. Days of Future Past takes advantage of an optimization model and HTTP/3 features, including (i) stream multiplexing, and (ii) request cancellation. We design a Mixed Integer Linear Programming (MILP) model that determines the optimal video qualities of both next segment requests and the segments currently located in the buffer. If better qualities for buffered segments are found, the client will send corresponding HTTP GET requests to retrieve them. Multiple segments (i.e., re-transmitted segments) might be downloaded simultaneously to upgrade some buffered but not yet played segments to avoid quality decreases using the stream multiplexing feature of QUIC. HTTP/3’s request cancellation will be used in case retransmitted segments will arrive at the client after their playout time. The experimental results shows that our proposed method is able to improve the QoE by up to 33.9 %.

Keywords: HTTP/3, QUIC, Days of Future Past, HAS, QoE

On October 27th, 2021, Negin Ghamsarian successfully defended her thesis on “Deep-Learning-Assisted Analysis of Cataract Surgery VIdeos” under the supervision of Prof. Klaus Schöffmann. The defense was chaired by Prof. Hermann Hellwagner and the examiners were Prof. Henning Müller (University of Applied Sciences Western Switzerland and the University of Geneva) and Prof. Raphael Sznitman (University of Bern). Congratulations to Dr. Ghamsarian for this great achievement!