23.October Tel-Aviv, Israel

AIM 2022

Advances in Image Manipulation workshop

in conjunction with ECCV 2022

Sponsors (TBA)

Call for papers

Image manipulation is a key computer vision tasks, aiming at the restoration of degraded image content, the filling in of missing information, or the needed transformation and/or manipulation to achieve a desired target (with respect to perceptual quality, contents, or performance of apps working on such images). Recent years have witnessed an increased interest from the vision and graphics communities in these fundamental topics of research. Not only has there been a constantly growing flow of related papers, but also substantial progress has been achieved.

Each step forward eases the use of images by people or computers for the fulfillment of further tasks, as image manipulation serves as an important frontend. Not surprisingly then, there is an ever growing range of applications in fields such as surveillance, the automotive industry, electronics, remote sensing, or medical image analysis etc. The emergence and ubiquitous use of mobile and wearable devices offer another fertile ground for additional applications and faster methods.

This workshop aims to provide an overview of the new trends and advances in those areas. Moreover, it will offer an opportunity for academic and industrial attendees to interact and explore collaborations.

This workshop builds upon the success of Advances in Image Manipulation (AIM) workshop at ICCV 2021, ECCV 2020,ICCV 2019, Mobile AI (MAI) workshop at CVPR 2022 , CVPR 2021 , Perceptual Image Restoration and Manipulation (PIRM) workshop at ECCV 2018 , the workshop and Challenge on Learned Image Compression (CLIC) editions at CVPR 2018, CVPR 2019, CVPR 2020, CVPR 2021, , CVPR 2022 and the New Trends in Image Restoration and Enhancement (NTIRE) editions: at CVPR 2017 , 2018, 2019 , 2020 , 2021 and 2022 and at ACCV 2016. Moreover, it relies on the people associated with the PIRM, CLIC, MAI, AIM, and NTIRE events such as organizers, PC members, distinguished speakers, authors of published papers, challenge participants and winning teams.

Papers addressing topics related to image/video manipulation, restoration and enhancement are invited. The topics include, but are not limited to:

  • Image-to-image translation
  • Video-to-video translation
  • Image/video manipulation
  • Perceptual manipulation
  • Image/video generation and hallucination
  • Image/video quality assessment
  • Image/video semantic segmentation
  • Perceptual enhancement
  • Multimodal translation
  • Depth estimation
  • Saliency and gaze estimation
  • Image/video inpainting
  • Image/video deblurring
  • Image/video denoising
  • Image/video upsampling and super-resolution
  • Image/video filtering
  • Image/video de-hazing, de-raining, de-snowing, etc.
  • Demosaicing
  • Image/video compression
  • Removal of artifacts, shadows, glare and reflections, etc.
  • Image/video enhancement: brightening, color adjustment, sharpening, etc.
  • Style transfer
  • Hyperspectral imaging
  • Underwater imaging
  • Aerial and satellite imaging
  • Methods robust to changing weather conditions / adverse outdoor conditions
  • Image/video manipulation on mobile devices
  • Image/video restoration and enhancement on mobile devices
  • Studies and applications of the above.

Important dates (TBU)

Challenges Event Date (always 23:59 CET)
Site online May 23, 2022
Release of train data and validation data May 24, 2022
Validation server online June 1, 2022
Final test data release, validation server closed July 23, 2022
Test phase submission deadline July 30, 2022
Fact sheets, code/executable submission deadline July 30, 2022
Preliminary test results release to the participants August 2, 2022
Paper submission deadline for entries from the challenges August 14, 2022 (EXTENDED)
Workshop Event Date (always 23:59 CET)
Paper submission deadline July 31, 2022 (EXTENDED)
Paper submission deadline (only for methods from AIM 2022 and Mobile AI 2022 challenges and papers reviewed elsewhere!) August 14, 2022 (EXTENDED)
Paper decision notification August 15, 2022
Camera ready deadline August 22, 2022
Workshop day October 23, 2022


Instructions and Policies
Format and paper length

A paper submission has to be in English, in pdf format, and at most 14 pages (excluding references) in single-column, ECCV style. The paper format must follow the same guidelines as for all ECCV 2022 submissions.
AIM 2022 and ECCV 2022 author guidelines

Double-blind review policy

The review process is double blind. Authors do not know the names of the chair/reviewers of their papers. Reviewers do not know the names of the authors.

Dual submission policy

Dual submission is not allowed. If a paper is submitted also to ECCV and accepted, the paper cannot be published both at the ECCV and the workshop.

Submission site



Accepted and presented papers will be published after the conference in ECCV Workshops proceedings together with the ECCV2022 main conference papers.

Author Kit

The author kit provides a LaTeX2e template for paper submissions.
Please refer to the example for detailed formatting instructions.


Organizers (TBU)

  • Radu Timofte, University of Wurzburg and ETH Zurich,
  • Andrey Ignatov, AI Benchmark and ETH Zurich,
  • Ren Yang, ETH Zurich,
  • Marcos V. Conde, University of Wurzburg,
  • Furkan Kınlı, Özyeğin University,

PC Members (TBU)

  • Mahmoud Afifi, Apple
  • Codruta Ancuti, UPT
  • Cosmin Ancuti, Polytechnic University of Timisoara
  • Boaz Arad, Ben-Gurion University of the Negev
  • Siavash Arjomand Bigdeli, DTU
  • Nick Barnes, Australian National University
  • Michael S. Brown, York University
  • Jianrui Cai, The Hong Kong Polytechnic University
  • Chia-Ming Cheng, MediaTek
  • Cheng-Ming Chiang, MediaTek
  • Sunghyun Cho, Samsung
  • Martin Danelljan, ETH Zurich
  • Christophe De Vleeschouwer, Université Catholique de Louvain
  • Tali Dekel, Weizmann Institute of Science
  • Chao Dong, SIAT
  • Weisheng Dong, Xidian University
  • Touradj Ebrahimi, EPFL
  • Paolo Favaro, University of Bern
  • Graham Finlayson, University of East Anglia
  • Corneliu Florea, University Politechnica of Bucharest
  • Peter Gehler, Amazon
  • Bastian Goldluecke, University of Konstanz
  • Shuhang Gu, OPPO & University of Sydney
  • Christine Guillemot, INRIA
  • Felix Heide, Princeton University & Algolux
  • Chiu Man Ho, OPPO,
  • Hiroto Honda, Mobility Technologies Co Ltd.
  • Zhe Hu, Hikvision Research
  • Andrey Ignatov, ETH Zurich
  • Sing Bing Kang, Zillow Group
  • Aggelos Katsaggelos, Northwestern University
  • Jan Kautz, NVIDIA
  • Vivek Kwatra, Google
  • Samuli Laine, NVIDIA
  • Jean-Francois Lalonde, Laval University
  • Christian Ledig, University of Bamberg
  • Seungyong Lee, POSTECH
  • Suyoung Lee, Seoul National University
  • Kyoung Mu Lee, Seoul National University
  • Victor Lempitsky, Skoltech & Samsung
  • Ales Leonardis, Huawei Noah's Ark Lab & University of Birmingham
  • Juncheng Li, The Chinese University of Hong Kong
  • Yawei Li, ETH Zurich
  • Stephen Lin, Microsoft Research
  • Ming-Yu Liu, NVIDIA Research
  • Chen Change Loy, Chinese University of Hong Kong
  • Guo Lu, Beijing Institute of Technology
  • Vladimir Lukin, National Aerospace University
  • Kede Ma, City University of Hong Kong
  • Vasile Manta, Technical University of Iasi
  • Rafal Mantiuk, University of Cambridge
  • Zibo Meng, OPPO
  • Rafael Molina, University of Granada
  • Yusuke Monno, Tokyo Institute of Technology
  • Hajime Nagahara, Osaka University
  • Vinay P. Namboodiri, IIT Kanpur
  • Federico Perazzi, Bending Spoons
  • Fatih Porikli, Qualcomm CR&D
  • Wenqi Ren, Chinese Academy of Sciences
  • Antonio Robles-Kelly, Deakin University
  • Andres Romero, ETH Zurich
  • Aline Roumy, INRIA
  • Yoichi Sato, University of Tokyo
  • Yoav Y. Schechner, Technion, Israel
  • Christopher Schroers, Disney Research | Studios
  • Nicu Sebe, University of Trento
  • Eli Shechtman, Creative Intelligence Lab at Adobe Research
  • Gregory Slabaugh, Queen Mary University of London
  • Sabine Süsstrunk, EPFL
  • Yu-Wing Tai, Kuaishou Technology & HKUST
  • Robby T. Tan, Yale-NUS College
  • Masayuki Tanaka, Tokyo Institute of Technology
  • Hao Tang, ETH Zurich
  • Jean-Philippe Tarel, IFSTTAR, France
  • Christian Theobalt, MPI Informatik
  • Qi Tian, Huawei Cloud & AI
  • Radu Timofte, University of Wurzburg & ETH Zurich
  • George Toderici, Google
  • Luc Van Gool, ETH Zurich & KU Leuven
  • Jue Wang, Tencent
  • Longguang Wang, National University of Defense Technology
  • Oliver Wang, Adobe Systems Inc
  • Ting-Chun Wang, NVIDIA
  • Yingqian Wang, National University of Defense Technology
  • Gordon Wetzstein, Stanford University
  • Ming-Hsuan Yang, University of California at Merced & Google
  • Ren Yang, ETH Zurich
  • Wenjun Zeng, Microsoft Research
  • Kai Zhang, ETH Zurich
  • Richard Zhang, UC Berkeley & Adobe Research
  • Yulun Zhang, ETH Zurich
  • Ruofan Zhou, EPFL
  • Jun-Yan Zhu, Carnegie Mellon University
  • Wangmeng Zuo, Harbin Institute of Technology

Invited Talks (TBU)

All the accepted AIM workshop papers have also oral presentation.
All the accepted AIM workshop papers are published under the book title "European Conference on Computer Vision Workshop (ECCVW)" by