Workshops

Submissions should be made through Microsoft CMT.

Paper submission deadline: 13 March 2020

Paper acceptance notification: 15 April 2020

Camera-ready paper submission deadline: 29 April 2020

Workshop Chairs



IEEE International Workshop of Artificial Intelligence in Sports (AI-Sports)

Organisers:

  • Prof. Huang-Chia Shih, Yuan Ze University, Taiwan
  • Prof. Rainer Lienhart, Augsburg University, Germany
  • Prof. Takahiro Ogawa, Hokkaido University, Japan
  • Prof. Jenq-Neng Hwang, University of Washington, USA

Description:

Sports data contains enormous potential in revolutionizing the sports industry. Coaches and teams are constantly searching for competitive sports data analytics that utilize AI and computer vision techniques to understand the deeper and hidden semantics of sports. By learning detailed statistics, coaches can assess defensive athletic performance and develop improved strategies. Sports data analytics is the process of analysing spatiotemporal content and sensor data from sports matches in online and offline scenarios. Currently, machine learning is already widely used in the sports industry. Many approaches have been proposed to extract semantic concepts or abstract attributes, such as objects, events, scene types, and captions, from sports videos. However, a limitation of conventional sports data analytics is that the domain-specific model can only be applied to analyse a single sport.

The goal of this workshop is to advance the field of research on the techniques of AI for sports data, develop more techniques to accurately evaluate and organize the data, and further strengthen the synergy between sports and science. Papers about machine learning, vision processing, and data sciences in sports and new forms of sports technologies are encouraged for submission.

Scope and Topics:

Topics of interest include, but are not limited to:

  • Object detection/modelling/recognition in sports data
  • Athletes motion capturing with learning algorithm in sports
  • Activities/actions recognition in sports data
  • 3D Sports and AR/VR
  • Artificial Intelligence strategy for sports
  • Tracking trajectories analysis with learning algorithm in sports
  • Semantic analysis in sports data
  • Tactics analysis for sports
  • Athletes’ decision-making
  • Supervised/unsupervised/reinforcement learning for sports data
  • Efficient learning algorithm for sports data compression
  • Energy- and resource-efficient machine learning architectures for large-scale sports data analytics
  • Sports video content analysis in the media cloud
  • Performance assessment in sports
  • Emerging applications of deep learning in sports content search, retrieval, recommendation, understanding, and summarization
  • Future trends and challenges for sports data analytics
  • New learning theories and models for sports data analysis and understanding
  • Other learning techniques from examples such as imitation learning and emerging cognition system in sports
  • New sports database and metrics to evaluate the benefit of sports analytics system
  • Survey papers regarding the topic of sports data analytics

Format:

Oral, posters, demos; half day.

Sponsors:

Ministry of Science and Technology Taiwan

Chinese Image Processing Pattern Recognition Society

Institute of Information & Computing Machinery, Taiwan.



AVS3 Video Coding Standard

Organisers:

  • Siwei Ma, Peking University, China
  • Lu Yu, Zhejiang University, China
  • Xiaozhen Zheng, DJI, China
  • Li Zhang, Bytedance, USA

Description:

AVS3 is the latest video coding standard developed by China AVS workgroup, targeting the emerging 4K/8K and VR applications. Till now, AVS3 has adopted many new efficient video coding tools, such as extended quad-tree block partitions, boundary filter for intra prediction, and flexible reference picture list management scheme. AVS3 shows significant coding gain over the previous video coding standards. Recently, Hisilicon has announced the first AVS3 8K@120p decoder chip at IBC2019. Moreover, AVS3 also did much exploration work in deep learning-based compression, where both piecemeal and end-to-end approaches were studied.

This workshop aims at bring together academic researchers, industrial practitioners, and individuals working on this emerging exciting research area to disseminate their new ideas, latest findings, and state-of-the-art results related to AVS3 developments.

Scope and Topics:

Topics of interest include, but are not limited to:

  • Coding tools
  • Software/hardware implementations
  • System transport
  • Quality evaluation
  • Learning based image/video compression

Format:

Oral, demos; half day.



The 2nd International Workshop on Big Surveillance Data Analysis and Processing (BIG-Surv)

Organisers:

  • John See, Multimedia University, Malaysia
  • Weiyao Lin, Shanghai Jiao Tong University, China
  • Xiatian Zhu, Samsung AI Centre, UK

Description:

With the rapid growth of video surveillance applications and services, the amount of surveillance videos has become extremely “big” which makes human monitoring tedious and difficult. Therefore, there exists a huge demand for smart surveillance techniques which can perform monitoring in an automatic or semi-automatic way. A number of challenges have arisen in the area of big surveillance data analysis and processing. Firstly, with the huge amount of surveillance videos in storage, video analysis tasks such as event detection, action recognition, and video summarization are of increasing importance in applications including events-of-interest retrieval and abnormality detection. Secondly, semantic data (e.g. objects’ trajectory and bounding boxes) has become an essential data type in surveillance systems owing much to the growth of its size and complexity, hence introducing new challenging topics, such as efficient semantic data processing and compression, to the community. Thirdly, with the rapid growth from the static centric-based processing to the dynamic computing among distributed video processing nodes/cameras, new challenges such as multi-camera analysis, person re-identification, or distributed video processing are being issued in front of us. To meet these challenges, there is great need to extend existing approaches or explore new feasible techniques.

Scope and Topics:

Topics of interest include, but are not limited to:

  • Event detection, action recognition, and activity analysis in surveillance videos
  • Multi-camera analysis and recognition
  • Object detection and tracking in surveillance videos
  • Recognition and parsing of crowded scenes
  • Person or group re-identification
  • Summarization and synopsis on surveillance videos
  • Surveillance scene parsing, segmentation, and analysis
  • Semantic data processing in large-scale surveillance systems
  • Data compression in surveillance systems
  • Robust face recognition and detection under low-resolution surveillance videos
  • Restoration and enhancement of degradations in low-quality surveillance videos

Format:

Oral and posters; half day.



The 1st ICME Workshop on Hyper-Realistic Multimedia for Enhanced Quality of Experience

Organisers:

  • Frédéric Dufaux, CNRS, France
  • Homer Chen, National Taiwan University, Taiwan
  • Ivan V. Bajić , Simon Fraser University, Canada
  • Søren Forchhammer, Technical University of Denmark, Denmark
  • Xiaolin Wu, McMaster University, Canada

Technical Programme Committee:

  • Anthony Vetro, MERL, USA
  • Atanas Gotchev, Tampere University, Finland
  • Dong Tian, InterDigital, USA
  • Fernando Pereira, Instituto Superior Técnico, Portugal
  • Jiaying Liu, Peking University, China
  • Joachim Keinert, IIS Fraunhofer, Germany
  • Mylene Farias, University of Brasília, Brasil
  • Patrick le Callet, University of Nantes, France
  • Peter Schelkens, VUB, Belgium
  • Rafal Mantiuk, University of Cambridge, UK
  • Sanghoon Lee, Yonsei University, Korea
  • Søren Bech, Bang & Olufsen, Denmark
  • Yonggang Wen, Nanyang Technological University, Singapore

Description:

The aim of hyper-realistic media is to faithfully represent the physical world. The ultimate goal is to create an experience, which is perceptually indistinguishable from a real scene. Traditional technologies can only capture a fraction of the audio-visual information, limiting the realism of the experience. Recent innovations in computers and audio-visual technology have made it possible to circumvent these bottlenecks in audio-visual systems. As a result, new multimedia signal processing areas have emerged such as light fields, ultra-high definition, high frame rate, high dynamic range imaging and novel 3D audio and sound field technologies. The novel combinations of those technologies can facilitate a hyper-realistic media experience. Without a doubt, this will be the future frontier for new multimedia systems. However, several technological barriers and challenges need to be overcome in developing the best solutions perceptually.

This first ICME workshop on Hyper-Realistic Multimedia for Enhanced Quality of Experience aims at bringing forward recent advances related to capturing, processing, and rendering technologies. The goal is to gather researchers with diverse and interdisciplinary backgrounds to cover the full multimedia signal chain, to efficiently develop truly perceptually enhanced multimedia systems.

Scope and Topics:

We seek unpublished high-quality papers within, but not limited to, the following topics:

  • Lightfield, point-cloud, volumetric imaging
  • High Dynamic Range imaging, Wide Color Gamut, Ultra High Definition
  • Multichannel, 3D audio and sound field systems, audio rendering
  • Hyper-realistic display technologies
  • Human perception modeling, perceptually-inspired processing
  • Processing and coding of hyper-realistic multimedia content
  • Subjective and objective quality assessment
  • Quality of experience
  • Hyper-realism and immersiveness
  • Human vision, clinical and experimental psychology and psychophysics

Format:

Oral, posters and panel; full day



The 1st International Workshop on Interactive Multimedia Retrieval

Organisers:

  • Werner Bailer, Joanneum Research, Austria
  • Klaus Schoeffmann, Klagenfurt University, Austria
  • Luca Rossetto, University of Zurich, Switzerland
  • Jakub Lokoč, Charles University, Czech Republic

Description:

With the recent increase in both volume and diversity of multimedia data, effective browsing and retrieval methods become increasingly important in order to deal with the available data and find the relevant documents. While this problem is well understood for textual documents, where an information need can often be expressed in sufficient detail with a textual query, the effective search in multimedia documents is generally more difficult.

The 1st International Workshop on Interactive Multimedia Retrieval calls for submissions related to interactive retrieval in and across all types of multimedia content.

Scope and Topics:

We invite submissions reporting on current work done in the context of e.g., the Video Browser Showdown or the Lifelog Search Challenge, as well as interactive variants of solutions to TRECVID, MediaEval or similar tasks. Submissions should describe methods, but also insights and lessons learned from participating in such benchmarks. In this context, contributions related (but not limited) to the following topics are invited:

  • Interactive Retrieval Approaches and Methods
  • Browsing and Interactive Search User Interfaces
  • Multi-User Search, Retrieval and Browsing
  • Understanding User Behaviour and Information Needs
  • Cross/Multi-Modal Retrieval Methods
  • Datasets, Evaluation Metrics and Protocols
  • Multimedia Indexing Methods
  • Video Summarization Methods
  • Interactive Multimedia System Design and Architecture

Format:

Oral and posters; full day. All presenters are encouraged to also demonstrate their interactive retrieval solutions in a dedicated demo session.



Data-driven Just Noticeable Difference for Multimedia Communication

Organisers:

  • Prof. Yun Zhang, Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, China
  • Prof. Raouf Hamzaoui, De Montfort University, UK
  • Prof. C.-C. Jay Kuo, University of Southern California, USA
  • Prof. Dietmar Saupe, University of Konstanz, Germany

Sponsor:

  • SFB-TRR 161


Description:

The Picture-wise Just Noticeable Difference (PJND) for a given subject, image/video, and compression scheme is the smallest distortion that the subject can perceive when the image/video is compressed with this compression scheme. The PJND is normally determined with subjective quality assessment tests for a large population of viewers. Knowing the PJND statistics allows to reduce the bitrate without perceptual quality loss for the chosen proportion of the population. The workshop seeks papers proposing novel techniques to determine or predict the PJND statistics, as well as using these statistics for image/video processing, compression, and communication. While the focus of the workshop is on the PJND concept, contributions to the conventional JND approach where a JND threshold is computed at the pixel or subband level are also welcome provided the work is data driven.

Scope and Topics:

Topics of interest include, but are not limited to:

  • PJND/JND datasets for images, video, 3D video, omni-directional images/video, and point clouds
  • PJND/JND visual attributes related to contents, displays, the environment and the human visual system
  • Data-driven computational models for PJND/JND
  • Machine learning techniques to estimate the PJND/JND
  • Evaluation methods and metrics for JND/PJND models
  • PJND/JND concept, visual attributes, perception and prediction models
  • Data-driven PJND/JND models and their application to visual perception
  • PJND/JND models and their application to multimedia signal processing, compression and communication

Format:

Oral and posters; half day

Invited Talk

Visual Perception and JND Modelling: Progress & Challenges

Speaker:

Weisi Lin, Nanyang Technological University, Singapore

Abstract:

Just-noticeable difference (JNDs), as perceptual thresholds of visibility, determine the minimal amounts of change for differences to be sensed by the human being (e.g., 75% of a population), and play an important role both explicitly and implicitly in many applications. The measurement, formulation and computationally-modeling for JND are the prerequisite for user-centric designs for turning human perceptual limitation into meaningful system advantages. In this talk, a holistic view will be presented on visual JND research and practice: absolute and utility-oriented JNDs; pixel-, subband- and picture- based JNDs; conventional and data-driven JND estimation; databases and model evaluation. Other factors influencing JND include culture and personality, as will be also highlighted. JND modeling for visual signals (naturally captured, computer-generated or mixed ones) has attracted much research interests so far, while those for audio, haptics, olfaction and gestation are expected to attract increasing research interests toward true multimedia. Possible new directions are then to be discussed in order to advance the relevant research.

Bio:

Weisi Lin researches in image processing, perception-based signal modelling and assessment, video compression, and multimedia communication systems. In the said areas, he has published 240+ international journal papers and 260+ international conference papers, 9 patents, 9 book chapters, 2 authored books and 3 edited books, as well as excellent track record in leading and delivering more than 10 major funded projects (with over S$7m research funding). He earned his BSc and MSc from Sun Yat-Sen University, China, and Ph.D from King’s College, University of London. He had been the Lab Head, Visual Processing, Institute for Infocomm Research (I2R). He is a Professor in School of Computer Science and Engineering, Nanyang Technological University, where he served as the Associate Chair (Graduate Studies) in 2013-2014.
He is a Fellow of IEEE and IET, and an Honorary Fellow of Singapore Institute of Engineering Technologists. He has been awarded Highly Cited Researcher 2019 by Web of Science, and elected as a Distinguished Lecturer in both IEEE Circuits and Systems Society (2016-17) and Asia-Pacific Signal and Information Processing Association (2012-13), and given keynote/invited/tutorial/panel talks to 20+ international conferences during the past 10 years. He has been an Associate Editor for IEEE Trans. on Image Processing, IEEE Trans. on Circuits and Systems for Video Technology, IEEE Trans. on Multimedia, IEEE Signal Processing Letters, Quality and User Experience, and Journal of Visual Communication and Image Representation. He was also the Guest Editor for 7 special issues in international journals, and chaired the IEEE MMTC QoE Interest Group (2012-2014); he has been a Technical Program Chair for IEEE Int’l Conf. Multimedia and Expo (ICME 2013), International Workshop on Quality of Multimedia Experience (QoMEX 2014), International Packet Video Workshop (PV 2015), Pacific-Rim Conf. on Multimedia (PCM 2012) and IEEE Visual Communications and Image Processing (VCIP 2017). He believes that good theory is practical, and has delivered 10+ major systems and modules for industrial deployment with the technology developed.



The 7th IEEE International Workshop on Mobile Multimedia Computing (MMC 2020)

Organisers:

  • Tian Gan, Shandong University, China
  • Wen-Huang Cheng, National Chiao Tung University, Taiwan
  • Kai-Lung Hua, National Taiwan University of Science and Technology, Taiwan
  • Vladan Velisavljevic, University of Bedfordshire, UK

Description:

The intimate presence of mobile devices in our daily life, such as smartphones and various wearable gadgets like smart watches, has dramatically changed the way we connect with the world around us. Nowadays, in the era of the Internet-of-Things (IoT), these devices are further extended by smart sensors and actuators and amend multimedia devices with additional data and possibilities. With a growing number of powerful embedded mobile sensors like camera, microphone, GPS, gyroscope, accelerometer, digital compass, and proximity sensor, there is a variety of data available and hence enables new sensing applications across diverse research domains comprising mobile media analysis, mobile information retrieval, mobile computer vision, mobile social networks, mobile human-computer interaction, mobile entertainment, mobile gaming, mobile healthcare, mobile learning, and mobile advertising. Therefore, the workshop on Mobile Multimedia Computing (MMC 2018) aims to bring together researchers and professionals from worldwide academia and industry for showcasing, discussing, and reviewing the whole spectrum of technological opportunities, challenges, solutions, and emerging applications in mobile multimedia.

Scope and Topics:

Topics of interest include but are not limited to:

  • Ubiquitous computing on mobile and wearable devices
  • Mobile visual search
  • Action/gesture/object/speech recognition with mobile sensor
  • Multimedia data in the IoT
  • Computational photography on mobile devices
  • Mobile social signal processing
  • Human computer interaction with mobile and wearable devices
  • Mobile virtual and augmented reality
  • Mobile multimedia content adaptation and adaptive streaming
  • Mobile multimedia indexing and retrieval
  • Power saving issues of mobile multimedia computing
  • Multi-modal and multi-user mobile sensing
  • Personalization, privacy and security in mobile multimedia
  • 2D/3D computer vision on mobile devices
  • User behavior analysis of mobile multimedia applications
  • Multimedia Cloud Computing
  • Other topics related to mobile multimedia computing

Format:

Oral and posters; half day

Awards:

The MMC Best Paper Award will be granted to the best overall paper. The selection is based on the quality, originality, and clarity of the submission.



Multimedia Services and Technologies for Smart-Health (MUST-SH 2020)

Organisers:

  • M. Shamim Hossain, King Saud University, KSA
  • Stefan Goebel, KOM, TU Darmstadt, Germany

Steering Committee:

  • Changsheng Xu, Multimedia Computing Group, Chinese Academy of Sciences, China (co-chair)
  • Abdulmotaleb El Saddik, University of Ottawa, Ottawa, Canada (co-chair)
  • Min, Chen, Huazhong University of Science and Technology (HUST), China
  • Mohsen Guizani, EiC IEEE Network
  • Athanasios Vasilakos, Lulea University of Technology, Sweden

Technical Chair:

  • Susan Malaika, IBM, USA
  • Md. Abdur Rahman, UPM, Saudi Arabia

Program Committee:

  • Robert Istepanian, Kingston University, UK
  • Zheng Chang, University of Jyväskylä, Finland
  • Min, Chen, Huazhong HUST, China
  • Athanasios Vasilakos, Lulea University of Technology, Sweden
  • Tom Baranowski, Baylor College of Medicine, USA
  • Stefan Goebel, Multimedia Communications Lab (KOM), TU Darmstadt, Germany
  • Yin Zhang, Zhongnan University of Economics and Law, China
  • Syed M. Rahman, University of Hawai, USA
  • Biao Song, Kyung Hee University, South Korea
  • Mukaddim Pathan, Australian National University, Australia
  • Gamhewage Chaminda de Silva, University of Tokyo, Japan
  • Kaoru Sezaki, University of Tokyo, Japan
  • Manzur Morshed, Monash University, Australia
  • Edward Delp, Purdue University, USA
  • Majdi Rawashdeh, New York University, UAE
  • Muhammad Ghulam, CCIS, King Saud University, KSA
  • Abdur Rahman, SITE, University of Ottawa, Canada
  • Al-Sakib Khan Pathan, IIUM, Malaysia
  • Jorge Parra, Ikerlan-IK4, Spain
  • Nada Philip, Kingston University, UK
  • Md. Mehedi Masud, Taif University, KSA
  • Mehedi Hassan, Kyung Hee University, South Korea
  • Atif Shamim, King Abdullah University Of Science & Technology, KSA
  • Josef Wiemeyer, TU Darmstadt, Germany
  • Lennart Nacke, University of Saskatchewan, Canada
  • Anders Drachen, AGORA Informatics, Denmark
  • Georgios Yannakakis, IT University of Kopenhagen, Denmark
  • Simon McCallum, Gjøvik University College, Hedmark, Norway

Description:

Today multimedia services and technologies play an important role in providing and managing smart healthcare services to anyone, anywhere and anytime seamlessly. These services and technologies facilitate doctors and other health care professionals to have immediate access to smart-health information for efficient decision making as well as better treatment. Researchers are working in developing various multimedia tools, techniques and services to better support smart -health initiatives. In particular, works in smart-health record management, elderly health monitoring, real-time access of medical images and video are of great interest.

Scope and Topics:

This workshop aims to report high quality research on recent advances in various aspects of smart health, more specifically to the state-of-the-art approaches, methodologies and systems in the design, development, deployment and innovative use of multimedia services, tools and technologies for smart health care. Authors are solicited to submit complete unpublished papers in the following, but not limited to the following topics of interest:

  • Edge-Cloud for Smart Healthcare
  • Deep learning approach for smart healthcare
  • Explainable artificial intelligence (AI) technology for secured smart healthcare
  • Serious Games for health
  • Multimedia big data for health care applications
  • Adaptive exergames for health
  • Fuzzy Logic Approach for smart healthcare monitoring
  • Multimedia Enhanced Learning, Training & Simulation for Smart Health
  • Sensor and RFID technologies for Smart health
  • Cloud-based smart health Services
  • Resource allocation for Media Cloud-assisted health care
  • IoT-Cloud for Smart Healthcare
  • Wearable health monitoring
  • Smart health service management
  • Context-aware Smart -Health services and applications
  • Elderly health monitoring
  • Collaborative Smart Health
  • Haptics for Surgical/medical Systems
  • 5G Tactile Internet for Smart Health

Format:

Oral and posters; full day



3D Point Cloud Processing, Analysis, Compression, and Communication (PC-PACC)

Organisers:

  • Hui Yuan, Shandong University, China
  • Huanqiang Zeng, Huaqiao University, China
  • Philip A. Chou, Google, USA
  • Pascal Frossard, EPFL, Switzerland

Description:

The trend over the past decade towards computational imaging has enabled vast amounts of 3D data to be sensed using collections of sensors. At the same time, new types of displays have made it possible to view these 3D data in increasingly natural ways. This combination of trends is giving rise to the next generation of media beyond images, audio, and video: immersive media. Immersive media can be represented in various ways. One representation in particular – 3D point clouds – is becoming increasingly popular, in part because many of the computational imaging systems that capture immersive media are fundamentally digital systems that sample the natural world at discrete 3D points. The signals sampled at these points become attributes of the points, for example color, reflectance, transparency, normal direction, motion direction, and so forth.

The purpose of this workshop is to promote further research and understanding of 3D point clouds and their processing, analysis, compression, and communication, by providing a venue for the exchange and discussion of recent results.

Scope and Topics:

The technical issues covered by this workshop include, but are not limited to:

  • Efficient compression for 3D point clouds, e.g., novel prediction technologies, transform methods, rate-distortion optimization methods, etc.,
  • 3D point cloud processing based on modern signal processing theory, e.g., graph signal processing,
  • 3D point cloud-based computer vision tasks, e. g., visual tracking, object detection, semantic segmentation, and recognition,
  • High-reliability and low-delay transmission management optimization for 3D point cloud transmission, and
  • Artificial neural network-based 3D point cloud analysis.

Format:

Oral, half day



Workshop on ICT and Multimedia Tools for Migrants Inclusion in Host societies (WIMMIH2020)

Organisers:

  • Dr Petros Daras, Centre for Research and Technology Hellas (CERTH), Greece
  • Dr Nicholas Vretos, Centre for Research and Technology Hellas (CERTH), Greece
  • Prof. Federico Alvarez, Universidad Politecnica de Madrid (UPM), Spain
  • Dr Theodoros Semertzidis, Centre for Research and Technology Hellas (CERTH), Greece
  • Prof. Yuri Adrian Tijerino, Kwansei Gakuin University, Japan

Description:

Migrants integration into host societies pose many challenges in different levels. From job seeking to education enrolling and from asylum seeker to vulnerable refugees, the spectrum of different tools that can be created to assist these people and host authorities is vast. Multimedia ICT solutions have been devised in order to cope with many of these issues. Artificial Intelligence (AI) and Machine Learning (ML) tools have been used thus far to assist migrants and host authorities to provide better services for the benefit of both migrants and host societies. It is evident that migration flows rise due to regional conflicts and/or environmental conditions shifting and that new tools need to be researched and develop to assist societies towards a smooth integration of these people to the host societies. In this volatile, intercultural, landscape and with the need to support many different languages, illiteracy problems, and lack of technology skills, multimedia approaches that reduce the need of written communication appear to be the most effective ones.

The aim of this workshop is to call for a coordinated effort to understand the scenarios and the challenges emerging in ICT solutions for migrants’ inclusion into host societies with AI and ML based Multimedia tools, identify the key tasks and evaluate the current state of the art in the specific domain. Moreover, this workshop will try to showcase innovative ideas in the area that aid on the smooth integration of migrants into host societies and discuss further directions. We solicit manuscripts in all fields that explore the synergies of Multimedia tools and AI, ML, towards the assistance of migrants and host authorities for a smooth inclusion of the former in a host society.

Scope and Topics:

We believe the workshop will offer a timely collection of research updates to benefit the researchers and practitioners working in the broad fields ranging from computer vision, artificial intelligence and machine learning, with emphasis on multimedia related solutions. To this end, we solicit original research and survey papers addressing the topics listed below (but not limited to):

  • AI technologies for multimedia game-based skill assessment;
  • AI technologies for video and image-based migration flow analysis;
  • AI technologies for skill-job matching;
  • AI technologies for video and image-based migration flow prediction;
  • AI technologies for automatic administrative multimedia-based assistance;
  • AI technologies for multimedia based intercultural communication assistance;
  • Data analytics and demo systems for large scale job seeking services;
  • Migration related multimedia datasets and evaluation protocols;
  • AI-assisted or human-AI co-operated technologies for administrative multimedia-based assistance;
  • Emerging new applications in Multimedia ICT Tools for Migrant Inclusion in host societies

Format:

Oral, half day



Tools for Creating XR Media Experiences

Organisers:

  • Hannes Fassold, Joanneum Research, Austria
  • Dimitrios Zarpalas, Centre for Research & Technology, Greece
  • Pablo Cesar, Centrum Wiskunde & Informatica and Delft University of Technology, Netherlands
  • Mario Montagud, i2CAT & University of Valencia, Spain

Description:

Extended Reality (XR), which includes Virtual Reality (VR), Augmented Reality (AR) and mixed reality (MR), creates entirely new ways for consumers to experience the world around them and interact with it. Within the last few years, improvements in sensor technology and processing power have led to tremendous advances in all aspects of XR hardware, and due to economies of scale of the massively growing XR market these devices are available now at a reasonable price point. On the production side, powerful low-cost systems for capturing 3D objects and volumetric video and 360° videos make it possible to create budget VR/AR productions. The same applies to the consumption side, where VR headsets like the Oculus Go or Playstation VR provide a highly immersive VR experience which is affordable for everyone.

Unfortunately, the development of tools and technologies for authoring, processing and delivering interactive XR experiences is lagging considerably behind the hardware development, which is definitely a hurdle for the cost-effective mass production of appealing XR content and scenarios. Lack of content in turn hinders broader adoption and acceptance of XR technologies by the consumer. For all these aspects, new approaches and technologies are needed in order to overcome the specific challenges of XR content creation (multimodal data, non-linear interactive storytelling, annotation and metadata models, novel compression techniques, bandwidth requirements, etc.).

This workshop asks for original contributions on new approaches, technologies and tools for creating, processing and delivering interactive XR media (3D/CGI content/point clouds, 360° video, 3DoF+/6DoF video, volumetric video, spatial audio…).

Please click here for further information.

Scope and Topics:

Topics of particular interest include, but are not limited to:

  • Efficient XR content acquisition and representation.
  • Compression and delivery to various platforms (HMD, smartphones, SmartTV / HbbTV, Web, …)
  • Subjective and objective assessment of XR scenarios (content quality, experiences…).
  • Semantic understanding of XR content (depth estimation, semantic segmentation, object recognition, pose estimation, action recognition, audio analysis, etc.).
  • Automating the XR content authoring process (e.g. providing automatic content annotation / storytelling)
  • Authoring interactions and navigation aids (e.g., elements for moving in time and space, avatars)
  • Authoring accessible XR experiences (e.g. subtitles, audio description, audio subtitling, sign language, …)

Format:

Oral and posters; half day



Media-Rich Fake News (MedFake)

Organisers:

  • Pradeep K. Atrey, University at Albany, State University of New York (SUNY), USA
  • Nitin Khanna, Indian Institute of Technology, Gandhinagar, India
  • Nalini K. Ratha, IBM Thomas J. Watson Research Center, USA
  • Luisa Verdoliva, University Federico II of Naples, Italy
  • Christian von der Weth, National University of Singapore, Singapore

Description:

Fake news is a type of social hacking designed to change a reader’s point of view, the effect of which may lead them to change their opinion about an individual, an organization, or a belief, and make misinformed decisions. With the advent of multimedia editing tools, fake news typically contains multiple types of media such as text, image, video and audio. Media-rich fake news can be easily made to look like a real one. Further, fakes news is prone to abrupt dissemination through increasing accessibility of the internet and online social media outlets. Although there has been a significant progress in multimedia security and forensics research, the modern web and social media avenues for creation and sharing of multimedia content poses fresh challenges related to fake content identification and mitigation. This workshop aims to bring forward further advances in the area of fake multimedia in terms of its proactive identification and the prevention of spread of such content.

Scope and Topics:

We invite latest and high-quality papers presenting or addressing issues related to media-rick fake news, but not limited to:

  • Media-rich fake email detection and prevention.
  • Media-rich fake news identification over social media.
  • Media-rich fake news mitigation over social media.
  • Content management policy for news publishers.
  • Content filtering for web.
  • Impact and severity of fake content.
  • Secure models, policies and practices for safe content filtering.
  • Identification and credibility of the author and the publishing source of fake content.
  • Fake content alert mechanisms.

Format:

Oral, panels; full day.