Multimedia Communication

18th International Conference on Network and Service Management (CNSM 2022)

Thessaloniki, Greece | 31 October – 4 November 2022

Conference Website

Minh Nguyen (Alpen-Adria-Universität Klagenfurt, Austria), Babak Taraghi (Alpen-Adria-Universität Klagenfurt, Austria), Abdelhak Bentaleb (National University of Singapore, Singapore), Roger Zimmermann (National University of Singapore, Singapore), and Christian Timmerer (Alpen-Adria-Universität Klagenfurt, Austria)

Abstract: Considering network conditions, video content, and viewer device type/screen resolution to construct a bitrate ladder is necessary to deliver the best Quality of Experience (QoE).
A large-screen device like a TV needs a high bitrate with high resolution to provide good visual quality, whereas a small one like a phone requires a low bitrate with low resolution. In
addition, encoding high-quality levels at the server side while the network is unable to deliver them causes unnecessary cost for the content provider. Recently, the Common Media Client Data (CMCD) standard has been proposed, which defines the data that is collected at the client and sent to the server with its HTTP requests. This data is useful in log analysis, quality of service/experience monitoring and delivery improvements.

cadlad

 

In this paper, we introduce a CMCD-Aware per-Device bitrate LADder construction (CADLAD) that leverages CMCD to address the above issues. CADLAD comprises components at both client and server sides. The client calculates the top bitrate (tb) — a CMCD parameter to indicate the highest bitrate that can be rendered at the client — and sends it to the server together with its device type and screen resolution. The server decides on a suitable bitrate ladder, whose maximum bitrate and resolution are based on CMCD parameters, to the client device with the purpose of providing maximum QoE while minimizing delivered data. CADLAD has two versions to work in Video on
Demand (VoD) and live streaming scenarios. Our CADLAD is client agnostic; hence, it can work with any players and ABR algorithms at the client. The experimental results show that CADLAD is able to increase the QoE by 2.6x while saving 71% of delivered data, compared to an existing bitrate ladder of an available video dataset. We implement our idea within CAdViSE — an open-source testbed for reproducibility.

 

IEEE Global Communications Conference (GLOBECOM)

December 4-8, 2022 |Rio de Janeiro, Brazil
Conference Website

Authors: Reza Farahani (Alpen-Adria-Universität Klagenfurt, Austria), Abdelhak Bentaleb (National University of Singapore, Singapore), Ekrem Cetinkaya (Alpen-Adria-Universität Klagenfurt, Austria), Christian Timmerer (Alpen-Adria-Universität Klagenfurt, Austria), Roger Zimmermann (National University of Singapore, Singapore), and Hermann Hellwagner (Alpen-Adria-Universität Klagenfurt, Austria)

Abstract: a cost-effective, scalable, and flexible architecture that supports low latency and high-quality live video streaming is still a challenge for Over-The-Top (OTT) service providers. To cope with this issue, this paper leverages Peer-to-Peer (P2P), Content Delivery Network (CDN), edge computing, Network Function Virtualization (NFV), and distributed video transcoding paradigms to introduce a hybRId P2P-CDN arcHiTecture for livE video stReaming (RICHTER). We first introduce RICHTER’s multi-layer architecture and design an action tree that considers all feasible resources provided by peers, edge, and CDN servers for serving peer requests with minimum latency and maximum quality. We then formulate the problem as an optimization model executed at the edge of the network. We present an Online Learning (OL) approach that leverages an unsupervised Self Organizing Map (SOM) to (i) alleviate the time complexity issue of the optimization model and (ii) make it a suitable solution for large-scale scenarios by enabling decisions for groups of requests instead of for single requests. Finally, we implement the RICHTER framework, conduct our experiments on a large-scale cloud-based testbed including 350 HAS players, and compare its effectiveness with baseline systems. The experimental results illustrate that RICHTER outperforms baseline schemes in terms of users’ Quality of Experience (QoE), latency, and network utilization, by at least 59%, 39%, and 70%, respectively.

Vignesh V Menon

ACM Multimedia Conference – Doctoral Symposium Track

Lisbon, Portugal | 10-14 October 2022

Vignesh V Menon (Christian Doppler Laboratory ATHENA, Alpen-Adria-Universität Klagenfurt)

Abstract: Rapid growth in multimedia streaming traffic over the Internet motivates the research and further investigation of the video coding performance of such services in terms of speed and Quality of Experience (QoE). HTTP Adaptive Streaming (HAS) is today’s de-facto standard to deliver clients the highest possible video quality. In HAS, the same video content is encoded at multiple bitrates, resolutions, framerates, and coding formats called representations. This study aims to (i) provide fast and compression-efficient multi-bitrate, multi-resolution representations, (ii) provide fast and compression-efficient multi-codec representations, (iii) improve the encoding efficiency of Video on Demand (VoD) streaming using content-adaptive encoding optimizations, and (iv) provide encoding schemes with optimizations per-title for live streaming applications to decrease the storage or delivery costs or/and increase QoE.

The ideal video compression system for HAS envisioned in this doctoral study.

 

Hadi

Secure Reversible Data Hiding in Encrypted Images based on Classification Encryption Difference

IEEE 24th Workshop on MultiMedia Signal Processing (MMSP)

September 26-28, 2022 | Shanghai, China

Authors: Lingfeng Qu (Southwest Jiaotong University),  Hadi Amirpour (Alpen-Adria-Universität Klagenfurt),   Mohammad Ghanbari (University of Essex, UK)and Christian Timmerer (Alpen-Adria-Universität Klagenfurt), Hongjie He (Southwest Jiaotong University)


Abstract: This paper introduces an algorithm to improve the security, efficiency, and
embedding capacity of reversible data hiding in encrypted images (RDH-EI). It is based on classification encryption difference and adaptive fixed-length coding. Firstly, the prediction error image is obtained, the difference with a bin value greater than the encryption threshold in the difference histogram is found, and it is further modified to obtain the embedding threshold range. Then, under the condition of ensuring that the difference inside and outside the threshold range is not confused, the difference within the threshold is only scrambled, and the difference outside the threshold is scrambled and mod encrypted. After obtaining the encrypted image, an adaptive difference fixed-length coding method is proposed to encode and compress the differences within the threshold. The secret data is embedded in the multiple most significant bits of the encoded difference. Experimental results show that the embedding capacity of the proposed algorithm is improved compared with the state-of-the-art algorithm.

In July-August 2022, the ATHENA Christian Doppler Laboratory hosted four interns working on the following topics:

  • Fabio Zinner: A Study and Evaluation on HTTP Adaptive Video Streaming using Mininet
  • Moritz Pecher: Dataset Creation and HAS Basics
  • Per-Luca Thalmann: Codec-war: is it necessary? Welcome to the multi-codec world
  • Georg Kelih: Server Client Simulator for QoE with practical Implementation

At the end of their internships, they presented their works and achieved results, and received official certificates from the university. We believe the joint work with them was beneficial for both the laboratory and the interns. We would like to thank the interns for their genuine interest, productive work, and excellent feedback about our laboratory.

Fabio Zinner: In my four weeks, I had an amazingly practical and theoretical experience which is very important for my future practical and academic line of work! It was great and fascinating working with Python, Mininet, Linux, FFMpeg, Gpac, Iperf, etc. I really liked working with ATHENA, and the experience I gathered was exceptional. Also, I am very happy that I had Reza Farahani as my supervisor!

Per-Luca Thalmann:I really enjoyed my 4 weeks at ATHENA. At first, I had to read a lot of articles and papers to get a basic understanding of Video Codecs and encoding. As I started my Main Project, which evaluated the performance of modern codecs with different video complexities, I noticed that everything I had read before was useful to progress faster towards my end goal. After I got the results of my script, which ran for over a week, I also noticed some outcomes which were not expected. Basically, that older codecs get at some very specific settings higher Scores than their successor. Whenever I got stuck or had any questions, my supervisor, Vignesh, helped me. I did not only improve my technical knowledge, but I also got a lot of insights into how research works, what is the motivation of research and also about the process for scientific research.

Georg Kelih:I worked by Athena as an Intern for a month and got the tasks to build a simulator which simulates the server-client communication (ABR, bitrate ladder, resource allocation) and shows the results in a graph and a Server Client script where the server runs on the local host and the client requests segments and plays them using python-vlc
My daily routine was pretty chill, not only we had only 30 hours to work, but also the programming was quite fun and challenging. So my day looked something like this I stand up go to work play a game of round table soccer and then start to work start Visual Studio Code and write the code I thought about yesterday hope that it runs, but it shows you just a few error messages start debugging then notice that it’s already time to eat something and that I am hungry, eat something and find finally after your lunch the silly error I made think about the new implementation and better ways to solve something and then it is already time to go, so you go to the Strandbad to swim a round and then drive home. Something like this, my daily routine looked like. For me, I think it was a bit too chill for my taste because I like the stress of a 40-hour week especially when I only work during my holidays.
But the rest was absolutely nice, especially that here by Athena are so many people from different countries is pretty cool. For myself, I learned not many new skills, but I found out about many new Linux tools and how to find information even more efficiently.

 

Hadi

Between Two and Six? Towards Correct Estimation of JND Step Sizes for VMAF-based Bitrate Laddering

14th International Conference on Quality of Multimedia Experience (QoMEX)
September 5-7, 2022 | Lippstadt, Germany

Hadi Amirpour (Alpen-Adria-Universität Klagenfurt)Raimund Schatz (AIT Austrian Institute of Technology, Austria)and Christian Timmerer (Alpen-Adria-Universität Klagenfurt)

Abstract: We currently witness the rapidly growing importance of intelligent video streaming quality optimization and reduction of video delivery costs. Per-Title encoding, in contrast to a fixed bitrate ladder, shows significant promise to deliver higher quality video streams by addressing the trade-off between compression efficiency and video characteristics such as resolution and frame rate. Selecting encodings with noticeable quality differences in between prevents the construction of an inefficient bitrate ladder that suffers from too similar quality representations. In this respect, the VMAF metric represents a promising foundation for bitrate laddering, as it currently yields the highest video quality prediction performance. However, the minimum noticeable quality difference, referred as to just-noticeable-difference (JND), has not been properly validated for VMAF yet, with existing sources proposing highly diverse ΔVMAF step sizes ranging from two to six.

Hadi

FuRA: Fully Random Access Light Field Image Compression

10th European Workshop on Visual Information Processing (EUVIP)
September 11-14, 2022 | Lisbon, Portugal

Hadi Amirpour (Alpen-Adria-Universität Klagenfurt),  Christine Guillemot (INRIA, France)and Christian Timmerer (Alpen-Adria-Universität Klagenfurt)

Abstract: Light fields are typically represented by multi-view images and enable post-capture actions such as refocusing and perspective shift. To compress a light field image, its view images are typically converted into a pseudo video sequence (PVS) and the generated PVS is compressed using a video codec. However, when using the inter-coding tool of a video codec to exploit the redundancy among view images, the possibility to randomly access any view image is lost. On the other hand, when video codecs independently encode view images using the intra-coding tool, random access to view images is enabled, however, at the expense of a significant drop in the compression efficiency. To address this trade-off, we propose to use neural representations to represent 4D light fields. For each light field, a multi-layer perceptron (MLP) is trained to map the light field four dimensions to the color space, thus enabling random access even to pixels. To achieve higher compression efficiency, neural network compression techniques are deployed. The proposed method outperforms the compression efficiency of HEVC inter-coding, while providing random access to view images and even pixel values.

Fully Random Access Light Field Image Compression

Low Latency Live Streaming Implementation in DASH and HLS

ACM Multimedia Conference – OSS Track

Lisbon, Portugal | 10-14 October 2022

Abdelhak Bentaleb (National University of Singapore), Zhengdao Zhan (National University of Singapore), Farzad Tashtarian (Christian Doppler Laboratory ATHENA, Alpen-Adria-Universität Klagenfurt), May Lim (National University of Singapore), Saad Harous (University of Sharjah), Christian Timmerer (Christian Doppler Laboratory ATHENA, Alpen-Adria-Universität Klagenfurt), Hermann Hellwagner (Christian Doppler Laboratory ATHENA, Alpen-Adria-Universität Klagenfurt), and Roger Zimmermann (National University of Singapore)

Low latency live streaming over HTTP using Dynamic Adaptive Streaming over HTTP (LL-DASH) and HTTP Live Streaming} (LL-HLS) has emerged as a new way to deliver live content with respectable video quality and short end-to-end latency. Satisfying these requirements while maintaining viewer experience in practice is challenging, and adopting conventional adaptive bitrate (ABR) schemes directly to do so will not work. Therefore, recent solutions including LoL$^+$, L2A, Stallion, and Llama re-think conventional ABR schemes to support low-latency scenarios. These solutions have been integrated with dash.js  that support LL-DASH. However, their performance in LL-HLS remains in question. To bridge this gap, we implement and integrate existing LL-DASH ABR schemes in the hls.js video player which supports LL-HLS.
Moreover, a series of real-world trace-driven experiments have been conducted to check their efficiency under various network conditions including a comparison with results achieved for LL-DASH in dash.js.

IEEE 14th Image, Video, and Multidimensional Signal Processing Workshop (IVMSP 2022)

June 26-29, 2022 | Nafplio, Greece

Conference Website

[Video]

Ekrem Çetinkaya (Christian Doppler Laboratory ATHENA, Alpen-Adria-Universität Klagenfurt), Minh Nguyen (Christian Doppler Laboratory ATHENA, Alpen-Adria-Universität Klagenfurt), and Christian Timmerer (Christian Doppler LaboratoryATHENA, Alpen-Adria-Universität Klagenfurt)

 

Abstract: Video is now an essential part of the Internet. The increasing popularity of video streaming on mobile devices and the improvement in mobile displays brought together challenges to meet user expectations. Advancements in deep neural networks have seen successful applications on several computer vision tasks such as super-resolution (SR). Although DNN-based SR methods significantly improve over traditional methods, their computational complexity makes them challenging to apply on devices with limited power, such as smartphones. However, with the improvement in mobile hardware, especially GPUs, it is now possible to use DNN based solutions, though existing DNN based SR solutions are still too complex. This paper proposes LiDeR, a lightweight video SR network specifically tailored toward mobile devices. Experimental results show that LiDeR can achieve competitive SR performance with state-of-the-art networks while improving the execution speed significantly, i.e., 267 % for X4 upscaling and 353 % for X2 upscaling compared to ESPCN.

Keywords: Super-resolution, Mobile machine learning, Video super-resolution.

 

Detection and Localization of Video Transcoding From AVC to HEVC Based on Deep Representations of Decoded Frames and PU Maps

IEEE Transactions on Multimedia

Haichao Yao (Beijing Jiaotong University), Rongrong Ni (Beijing Jiaotong University), Hadi Amirpour (Alpen-Adria-Universität Klagenfurt), Christian Timmerer (Alpen-Adria-Universität Klagenfurt)Yao Zhao (Beijing Jiaotong University).