ICCV Logo
Search by typing & pressing enter

YOUR CART

The 11th IEEE International Workshop on

Analysis and Modeling of Faces and Gestures

In conjunction with ICCV 2023
October 2 starting from 8:30 AM (Local Time)

Call for Papers
We have experienced rapid advances in the face, gesture, and cross-modality (e.g., voice and face) technologies. This is thanks to deep learning (i.e., dating back to 2012, AlexNet) and large-scale, labeled datasets. The progress in deep learning continues to push renowned public databases to near saturation, thus calling for evermore challenging image collections to be compiled as databases. In practice, and even widely in applied research, using off-the-shelf deep learning models has become the norm, as numerous pre-trained networks are available for download and are readily deployed to new, unseen data (e.g., VGG-Face, ResNet). We have almost grown “spoiled” from such luxury, which, in all actuality, has enabled us to stay hidden from many truths. Theoretically, the truth behind what makes neural networks more discriminant than ever before is still, in all fairness, unclear. Rather, they act as a sort of black box to most practitioners and even researchers alike. More troublesome is the absence of tools to quantitatively and qualitatively characterize existing deep models, which could yield greater insights about these all-so-familiar black boxes. With the frontier moving forward at rates incomparable to any spurt of the past, challenges such as high variations in illumination, pose, age, etc., now confront us. However, state-of-the-art deep learning models often fail when faced with such challenges owing to the difficulties in modeling structured data and visual dynamics.

Alongside the effort spent on conventional face recognition is the research done across modality learning, such as face and voice, gestures in imagery, and video motion, along with several other tasks. This line of work has attracted attention from industry and academic researchers from all sorts of domains. Additionally, and in some cases with this, there has been a push to advance these technologies for social media-based applications. Regardless of the exact domain and purpose, the following capabilities must be satisfied: face and body tracking (e.g., facial expression analysis, face detection, gesture recognition), lip reading and voice understanding, face and body characterization (e.g., behavioral understanding, emotion recognition), face, body, and gesture characteristic analysis (e.g., gait, age, gender, ethnicity recognition), group understanding via social cues (e.g., kinship, non-blood relationships, personality), and visual sentiment analysis (e.g., temperament, arrangement). Thus, needing to be able to create effective models for visual certainty has significant value in both the scientific communities and the commercial market, with applications that span topics of human-computer interaction, social media analytics, video indexing, visual surveillance, and internet vision. Currently, researchers have made significant progress addressing many of these problems, especially when considering off-the-shelf and cost-efficient vision products available these days, e.g. Intel RealSense, SHORE, and Affdex. Nonetheless, serious challenges remain, which only amplify when considering the unconstrained imaging conditions captured by different sources focused on non-cooperative subjects. It is these latter challenges that especially grab our interest, as we sought to bring together cutting-edge techniques and recent advances in deep learning to solve the challenges in the wild.

This one-day serial workshop (AMFG2023) provides a forum for researchers to review the recent progress of recognition, analysis, and modeling of face, body, and gesture while embracing the most advanced deep learning systems available for face and gesture analysis, particularly under an unconstrained environment like social media and across modalities like face-to-voice. The workshop includes up to two keynotes and peer-reviewed papers (oral and poster). Original high-quality contributions are solicited on the following topics:
  • Novel deep model, deep learning survey, or comparative study for face/gesture recognition;
  • Data-driven or physics-based generative models for faces, poses, and gestures; deep learning for internet-scale soft biometrics and profiling: age, gender, ethnicity, personality, kinship, occupation, beauty ranking, and fashion classification by facial or body descriptor;
  • Deep learning for detection and recognition of faces and bodies with large 3D rotation, illumination change, partial occlusion, unknown/changing background, and aging (i.e., in the wild); especially large 3D rotation robust face and gesture recognition;
  • Motion analysis, tracking, and extraction of face and body models captured from several nonoverlapping views;
  • Face, gait, and action recognition in low-quality (e.g., blurred), or low-resolution video from fixed or mobile device cameras;
  • AutoML for face and gesture analysis;
  • Social/psychological-based studies that aid in understanding computational modeling and building better automated face and gesture systems with interactive features;
  • Multimedia learning models involving faces and gestures (e.g., voice, wearable IMUs, and face);
  • Trustworthy learning for face and gesture analysis, e.g., fairness, explainability and transparency;
  • Other applications involving face and gesture analysis.


Related Workshops
The first AMFG was held in conjunction with the 2003 ICCV in Nice, France. So far, it has been successfully held TEN times. The homepages of the last three AMFG workshops are as follows: 
  • AMFG2018@CVPR: https://web.northeastern.edu/smilelab/AMFG2018/
  • AMFG2019@CVPR: https://web.northeastern.edu/smilelab/amfg2019/
  • AMFG2021@CVPR: https://web.northeastern.edu/smilelab/amfg2021/
Face and gesture (hands) modeling have been long-standing problems in the computer vision community, and have also been widely explored and studied in many other workshops in recent years, yet with different emphases compared with AMFG as follows:
  • Face recognition towards security concerns, such as fairness in ChaLearn2020@ECCV, face antispoofing in ChaLearn2021@ICCV, and masked face recognition in MFR2021@ICCV
  • Hands modeling for action understanding, such as HANDS2022@ECCV and HBHA2022@ECCV
  • Face and gesture modeling in VR/AR, such as WCPA2022@ECCV and CV4ARVR2022@CVPR
The proposed workshop will focus on the fundamental research centering on face and gesture, and thus provide theoretical and technical support to the above applications. The topics covered by AMFG will also benefit the community in a more generalized context, including human-computer interaction, multimodal learning, egocentric vision, artificial ethics, robotics, etc.
Important Dates
[ 08/11/2023 ] Submission Deadline

[ 09/01/2023 ] Notification

[ 09/15/2023 ] Camera-Ready Due

Author Guidelines
Submissions are handled via the workshop's CMT website:    https://cmt3.research.microsoft.com/AMFG2023  


Following the guideline of ICCV2023: https://iccv2023.thecvf.com/submission.guidelines-361600-2-20-16.php

  • 8 pages (excluding references)

  • Anonymous

  • Using ICCV LaTex templates
Workshop Organizers
General Chair

Yun (Raymond) Fu, Northeastern University, USA.

http://www1.ece.neu.edu/~yunfu/

Workshop Chairs

Ming Shao, University of Massachusetts, Dartmouth, USA.

http://www.cis.umassd.edu/~mshao/

Sheng Li, University of Virginia, USA.

https://sheng-li.org/

Hongfu Liu, Brandeis University, USA.

http://hongfuliu.com/

Joseph P. Robinson, Northeastern University, Tufts University, USA.

https://www.jrobsvision.com/

Zhiqiang Tao, Rochester Institute of Technology, USA.

https://ztao.cc/

Yu Yin, Northeastern University, USA.

https://yin-yu.github.io/
Program Committee
  • Jake Aggarwal, UT Austin, USA
  • Marian Bartlett, UCSD, USA
  • Nadia Berthouze, UCL, UK
  • Aaron Bovick, GaTech, USA
  • Richard Bowden, Surrey, UK
  • Robert Collins, Penn State, USA
  • Tim Cootes, Manchester, UK
  • James Davis, Ohio State, USA
  • Fernando de la Torre, CMU, USA
  • Daniel Gatica Perez, IDIAP, Switzerland
  • Hatice Gunes, QMUL, UK
  • David Jacobs, Maryland, USA
  • Ron Kimmel, Technion, Israel
  • Josef Kittler, Surrey, UK
  • Aleix Martinez, Ohio State, USA
  • Vittorio Murino, Verona, Italy
  • Gerard Medioni, USC, USA
  • Alice O'Toole, UT Dalles, USA
  • Ioannis Patras, QMUL, UK
  • Matti Pietikainen, Oulu, Finland
  • Ian Reid, Oxford, UK
  • Marios Savvides, CMU, USA
  • Luc van Gool, ETHZ, Switzerland
  • Harry Wechler, GMU, USA
  • P. Wurtz, Bochum, Germany
  • Tao Xiang, QMUL, UK
  • Stefanos Zafeiriou, Imperial College, UK

Keynotes

Rama Chellappa (Tentative), Johns Hopkins University.

https://engineering.jhu.edu/faculty/rama-chellappa/

Title: TBD
Abstract. TBD.
Bio. TBD.


Matthew Turk (Tentative), Toyota Technological Institute at Chicago.

https://home.ttic.edu/~mturk/

Title: TBD
Abstract. TBD.
Bio. TBD.


Todd Zickler (Tentative), Harvard University.

http://www.eecs.harvard.edu/~zickler/Main/HomePage

Title: TBD
Abstract. TBD.
Bio. TBD.


Olga Russakovsky (Tentative), Princeton University.

https://www.cs.princeton.edu/~olgarus/

Title: TBD
Abstract. TBD.
Bio. TBD.



Program Schedule
on October 2 (Local Time)
8:30 AM
Chairs' opening remarks
8:45 AM
Invited talk I
9:30 AM
Coffee break I
10:00 AM
Oral session I
12:30 PM
Lunch break
2:00 PM
Invited talk II
2:45 PM
Coffee break II
3:15 PM
Oral session II
5:00 PM
Best Paper Announcement and Conclusion