Hadi

EUVIP 2022 Special Session on

“Machine Learning for Immersive Content Processing”

September, 2022, Lisbon, Portugal

Link

Organizers:

  • Hadi Amirpour, Klagenfurt University, Austria
  • Christine Guillemot, INSA, France
  • Christian Timmerer, Klagenfurt University, Austria

 

Brief description:

The importance of remote communication is becoming more and more important in particular after  COVID-19 crisis. However, to bring a more realistic visual experience, more than the traditional two-dimensional (2D) interfaces we know today is required. Immersive media such as 360-degree, light fields, point cloud, ultra-high-definition, high dynamic range, etc. can fill this gap. These modalities, however, face several challenges from capture to display. Learning-based solutions show great promise and significant performance in improving traditional solutions in addressing the challenges. In this special session, we will focus on research works aimed at extending and improving the use of learning-based architectures for immersive imaging technologies.

Important dates:

Paper Submissions: 6th June, 2022
Paper Notifications: 11th July, 2022

 

MPEG, specifically, ISO/IEC JTC 1/SC 29/WG 3 (MPEG Systems), has been just awarded a Technology & Engineering Emmy® Award for its ground-breaking MPEG-DASH standard. Dynamic Adaptive Streaming over HTTP (DASH) is the first international de-jure standard that enables efficient streaming of video over the Internet and it has changed the entire video streaming industry including — but not limited to —  on-demand, live, and low latency streaming and even for 5G and the next generation of hybrid broadcast-broadband. The first edition has been published in April 2012 and MPEG is currently working towards publishing the 5th edition demonstrating an active and lively ecosystem still being further developed and improved to address requirements and challenges for modern media transport applications and services.

This award belongs to 90+ researchers and engineers from around 60 companies all around the world who participated in the development of the MPEG-DASH standard for over 12 years.

From left to right: Kyung-mo Park, Cyril Concolato, Thomas Stockhammer, Yuriy Reznik, Alex Giladi, Mike Dolan, Iraj Sodagar, Ali Begen, Christian Timmerer, Gary Sullivan, Per Fröjdh, Young-Kwon Lim, Ye-Kui Wang. (Photo © Yuriy Reznik)

Christian Timmerer, director of the Christian Doppler Laboratory ATHENA, chaired the evaluation of responses to the call for proposals and since that served as MPEG-DASH Ad-hoc Group (AHG) / Break-out Group (BoG) co-chair as well as co-editor for Part 2 of the standard. For a more detailed history of the MPEG-DASH standard, the interested reader is referred to Christian Timmerer’s blog post “HTTP Streaming of MPEG Media” (capturing the development of the first edition) and Nicolas Weill’s blog post “MPEG-DASH: The ABR Esperanto” (DASH timeline).

The 13th ACM Multimedia Systems Conference (ACM MMSys 2022)

June 14–17, 2022 |  Athlone, Ireland

Conference Website

Reza Shokri Kalan (Digiturk Company, Istanbul), Reza Farahani (Alpen-Adria-Universität Klagenfurt), Emre Karsli (Digiturk Company, Istanbul), Christian Timmerer (Alpen-Adria-Universität Klagenfurt), and Hermann Hellwagner (Alpen-Adria-Universität Klagenfurt)

Over-the-Top (OTT) service providers need faster, cheaper, and Digital Rights Management (DRM)-capable video streaming solutions. Recently, HTTP Adaptive Streaming (HAS) has become the dominant video delivery technology over the Internet. In HAS, videos are split into short intervals called segments, and each segment is encoded at various qualities/bitrates (i.e., representations) to adapt to the available bandwidth. Utilizing different HAS-based technologies with various segment formats imposes extra cost, complexity, and latency to the video delivery system. Enabling an integrated format for transmitting and storing segments at Content Delivery Network (CDN) servers can alleviate the aforementioned issues. To this end, MPEG Common Media Application Format (CMAF) is presented as a standard format for cost-effective and low latency streaming. However, CMAF has not been adopted by video streaming providers yet and it is incompatible with most legacy end-user players. This paper reveals some useful steps for achieving low latency live video streaming that can be implemented for non-DRM sensitive contents before jumping to CMAF technology. We first design and instantiate our testbed in a real OTT provider environment, including a heterogeneous network and clients, and then investigate the impact of changing format, segment duration, and Digital Video Recording (DVR) window length on a real live event. The results illustrate that replacing the transport stream (.ts) format with fragmented MP4 (.fMP4) and shortening segments’ duration reduces live latency significantly.

 

 

 

 

 

 

 

 

Keywords: HAS, DASH, HLS, CMAF, Live Streaming, Low Latency

 

Hadi

colocated with ACM Multimedia 2022

October, 2022, Lisbon, Portugal

Workshop Chairs:

  • Irene Viola, CWI, Netherlands
  • Hadi Amirpour, Klagenfurt University, Austria
  • Asim Hameed, NTNU, Norway
  • Maria Torres Vega, Ghent University, Belgium

Topics of interest include, but are not limited to:

  • Novel low latency encoding techniques for interactive XR applications
  • Novel networking systems and protocols to enable interactive immersive applications. This includes optimizations ranging from hardware (i.e., millimeter-wave networks or optical wireless), physical and MAC layer up to the network, transport and application layers (such as over the top protocols);
  • Significative advances and optimization in 3D modeling pipelines for AR/VR visualization, accessible and inclusive GUI, interactive 3D models;
  • Compression and delivery strategies for immersive media contents, such as omnidirectional video, light fields, point clouds, dynamic and time varying meshes;
  • Quality of Experience management of interactive immersive media applications;
  • Novel rendering techniques to enhance interactivity of XR applications;
  • Application of interactive XR to different areas of society, such as health (i.e., virtual reality exposure therapy), industry (Industry 4.0), XR e-learning (according to new global aims);

Dates:

  • Submission deadline: 20 June 2022, 23:59 AoE
  • Notifications of acceptance: 29 July 2022
  • Camera ready submission: 21 August 2022
  • Workshop: 10th or 14th October

ALIS’22: Artificial Intelligence for Live Video Streaming


colocated with ACM Multimedia 2022


October 2022, Lisbon, Portugal

Download ALIS’22 Poster/ CfP

 

Project Lead: H. Hellwagner, Ch. Timmerer

Abstract: Immersive telepresence technologies will have game-changing impacts on interactions amongst individuals or with non-human objects (e.g. machines), in cyberspace with blurred boundaries between the virtual and physical world. The impacts of this technology are expected to range in a variety of vertical sectors, including education and training, entertainment, healthcare, manufacturing industry, etc. The key challenges include limitations of both the application platform and the underlying network support to achieve seamless presentation, processing, and delivery of immersive telepresence content at a large scale. Innovative design, rigorous validation, and testing exercises aim to fulfill the key technical requirements identified such as low-latency communication, high bandwidth demand, and complex content encoding/rendering tasks in real-time. The industry-leading SPIRIT consortium will build on the existing TRL4 application platforms and network infrastructures developed by the project partners, aiming to address key technical challenges and further develop all major aspects of telepresence technologies to achieve targeted TRL7. The SPIRIT Project will focus its innovations in network-layer, transport-layer, application/content-layer techniques, as well as security and privacy mechanisms to facilitate the large-scale operation of telepresence applications. The project team will develop a fully distributed, interconnected testing infrastructure across two geographical sites in Germany and UK, allowing large-scale testing of heterogeneous telepresence applications in real-life Internet environments. The network infrastructure will host two mainstream application
environments based on WebRTC and low-latency DASH. In addition to the project-designated use case scenarios, the project team will test a variety of additional use cases covering heterogeneous vertical sectors through FSTP participation.

ACM Mile-High video 2022 (MHV)

March 01-03, 2022 | Denver, CO, USA

Conference Website

Authors: Minh Nguyen (Christian Doppler Laboratory ATHENA, Alpen-Adria-Universität Klagenfurt, Austria), Christian Timmerer (Christian Doppler Laboratory ATHENA, Alpen-Adria-Universität Klagenfurt, Austria), Stefan Pham (Fraunhofer FOKUS, Germany), Daniel Silhavy (Fraunhofer FOKUS, Germany), Ali C. Begen (Ozyegin University, Turkey)

Abstract: With the introduction of HTTP/3 (H3) and QUIC at its core, there is an expectation of significant improvements in Web-based secure object delivery. As HTTP is a central protocol to the current adaptive streaming methods in all major over-the-top (OTT) services, an important question is what H3 will bring to the table for such services. To answer this question, we present the new features of H3 and QUIC, and compare them to those of H/1.1/2 and TCP. We also share the latest research findings in this domain.

Keywords: HTTP adaptive streaming, QUIC, CDN, ABR, OTT, DASH, HLS.

The Emmy® Awards do not only honour the work of actors and directors, but they also recognise technologies that are steadily improving the viewing experience for consumers. This year, the winners include the MPEG DASH Standard. Christian Timmerer (Department of Information Technology) played a leading role in its development. Read more about it here.

ACM Mile-High video 2022 (MHV)

March 01-03, 2022 | Denver, CO, USA

Conference Website

Minh Nguyen (Christian Doppler Laboratory ATHENA, Alpen-Adria-Universität Klagenfurt), Ekrem Çetinkaya (Christian Doppler Laboratory ATHENA, Alpen-Adria-Universität Klagenfurt), Hermann Hellwagner (Christian Doppler Laboratory ATHENA, Alpen-Adria-Universität Klagenfurt), and Christian Timmerer (Christian Doppler Laboratory ATHENA, Alpen-Adria-Universität Klagenfurt)

Abstract: The advancement of mobile hardware in recent years made it possible to apply deep neural network (DNN) based approaches on mobile devices. This paper introduces a lightweight super-resolution (SR) network, namely SR-ABR Net, deployed at mobile devices to upgrade low-resolution/low-quality videos and a novel adaptive bitrate (ABR) algorithm, namely WISH-SR, that leverages SR networks at the client to improve the video quality depending on the client’s context. WISH-SR takes into account mobile device properties, video characteristics, and user preferences. Experimental results show that the proposed SR-ABR Net can improve the video quality compared to traditional SR approaches while running in real-time. Moreover, the proposed WISH-SR can significantly boost the visual quality of the delivered content while reducing both bandwidth consumption and the number of stalling events.

Keywords: Super-resolution, Deep Neural Networks, Mobile Devices, ABR

Hadi

On Tuesday the 25th of January 2022, Hadi Amirpour successfully defended his Ph.D. thesis under supervision of Assoc.-Prof. DI Dr. Christian Timmerer and Assoc.-Prof. Dr. Klaus Schöffmann. The defense was chaired by Assoc.-Prof. DI Dr. Mathias Lux and the examiners were Emeritus Prof. Dr. Mohammad Ghanbari (University of Essex, UK) and Univ.-Prof. DI Dr. Hermann Hellwagner (University of Klagenfurt).

We are pleased to congratulate Dr. Hadi Amirpour on passing his Ph.D. exam!