The 6th International Workshop on Affective Social Multimedia Computing (ASMMC 2021)
The 6th International Workshop on
Affective Social Multimedia Computing
October 18-22, 2021, Montreal, Canada
Call for Papers
Affective social multimedia computing is an emergent research topic for both affective computing and multimedia research communities. Social multimedia is fundamentally changing how we communicate, interact, and collaborate with other people in our daily lives. Comparing with well-organized broadcast news and professionally made videos such as commercials, TV shows, and movies, social multimedia media computing imposes great challenges to research communities. Social multimedia contains much affective information. Effective extraction of affective information from social multimedia can greatly help social multimedia computing (e.g., processing, index, retrieval, and understanding). Although much progress have been made in traditional multimedia research on multimedia content analysis, indexing, and retrieval based on subjective concepts such as emotion, aesthetics, and preference, affective social multimedia computing is a new research area. The affective social multimedia computing aims to proceed affective information from social multi-media. For massive and heterogeneous social media data, the research requires multidisciplinary understanding of content and perceptual cues from social multimedia. From the multimedia perspective, the research relies on the theoretical and technological findings in affective computing, machine learning, pattern recognition, signal/multimedia processing, computer vision, speech processing, behavior and social psychology. Affective analysis of social multimedia and interaction is attracting growing attention from industry and businesses that provide social networking sites, content-sharing services, distribute and host the media, social interaction with artificial agents. This workshop focuses on the analysis of affective signals in interaction (multimodal analyses enabling artificial agents in Human-Machine Interaction, social Interaction with artificial agents) and social multimedia (e.g., twitter, wechat, weibo, youtube, facebook, etc).
The 1st, 2nd, 3rd, 4th, 5th ASMMC workshop has been successfully held in Xi'an, China on September 21, 2015, Seattle, USA on July 15, 2016, Stockholm, Sweden on August 25, 2017, and Seoul, Korea on October 26, 2018, July 2, 2019 respectively. We take the 6th ASMMC to ACM ICMI 2021 come back again to Affective Computing & Intelligent Interaction for investigating affective computing technology to become available and accessible to education, health, transport, cities, home and entertainments.
It seeks contributions on various aspects of affective computing in interaction and social multimedia and interaction on related theory, methodology, algorithms, techniques, and applications
- Affective human-machine interaction or human-human interaction
- Affective/Emotional content analysis of images, videos, speech, music, metadata (text, symbols, etc.)
- Affective indexing, ranking, and retrieval on big social media data
- Affective computing in social multimedia by multimodal integration (face expression, gesture, posture, speech, text/language)
- Emotional implicit tagging and interactive systems
- User interests and behaviour modeling in social multimedia
- Video and image summarization based on affect
- Affective analysis of social media and harvesting the affective response of crowd
- Affective generation in social multimedia, expressive text-to-speech and expressive language translation
- Zero/One/Few-shot learning for emotion recognition
- Multimodal Analyses enabling Artificial Agents in Human-Machine Interaction
- Social Interaction with Artificial Agents
- Applications of affective social multimedia computing
Workshop General Chair
- Youjun XIONG
UBTech Robotics Corporate
Workshop Chairs
- Dongyan HUANG
Senior Member, IEEE, Principle Scientist, Text-to-Speech Synthesis Lab
Audio, Speech and Language Technology, UBTech Robotics Corp - Björn SCHULLER
Fellow of the IEEE, Senior Member ACM
Full Professor & Head of the ZD.B Chair of Embedded Intelligence for Health Care & Wellbeing, University of Augsburg, Germany
Reader (Associate Professor) in Machine Learning, Department of Computing, Imperial College London, London/U.K
Chief Executive Officer (CEO) and Co-Founder, audEERING GmbH, Gilching/Germany - Jianhua TAO
Full Professor & Deputy Director, National Laboratory of Pattern Recognition (NLPR)
Institute of Automation, Chinese Academy of Sciences - Lei XIE
Senior Member, IEEE
Dr./ Professor
Shannxi Provincial Key Lab of Speech and Image Information Processing (SAIIP)
School of Computer Science
Northwestern Polytechnical University, Xi'an, China - Jie YANG
Fellow, IEEE
Program director
Division of Information and Intelligent Systems
National Science Foundation (NSF) of USA
Organizing Committee
- Shih-Fu CHANG
Columbia University, USA - Stephen COX
University of East Anglia, UK - Minghui DONG
Institute for Infocomm Research, Singapore - Wolfgang HUERST
Utrecht University, Netherlands - Qiang JI
Rensselaer Polytechnic Institute, USA - Jia JIA
Tsinghua university, China - Qin JIN
Renmin University of China, China - Dongmei JIANG
Northwestern Polytechnical University, Xi'an, China - Bo LI
Google, USA - Haizhou LI
Institute for Infocomm Research, Singapore - Ming LI
Duke-Kunshan University, China - Jiebo LUO
University of Rochester, USA - Hichem SAHLI
Vrije Universiteit Brussel, Belgium - Vidhyasaharan SETHU
The University of New South Wales, Australia
- Rainer STIEFELHAGEN
Karlsruhe Institute of Technology, USA - Yan TONG
University of South Carolina, USA - Jia-Ching WANG
National Central University - Chung-Hsien WU
National Cheng Kung University, Taiwan - Zhiyong Wu
Tsinghua University, China - Zhizheng Wu
Edinburgh University, UK - Changsheng XU
Chinese Academy of Sciences, China - Shuicheng YAN
National University of Singapore, Singapore - Yanning ZHANG
Northwestern Polytechnical University, China - Peng ZHANG
Northwestern Polytechnical University, China - Yuexian ZOU
Peking University, China - Zhongfei ZHANG
Binghamton University, USA - Xuan ZHU
Samsung R&D Institute of China, China
Keynote
Qin Jin is a professor in School of Information at Renmin University of China (RUC), where she leads the AI·M³ lab. She received her Ph.D. degree in 2007 at Carnegie Mellon University. Before joining RUC in 2013, she was a research faculty (2007-2012) and a research scientist (2012) at Carnegie Mellon University and IBM China Research Lab respectively. Her research interests are in intelligent multimedia computing and human computer interaction. Her team’s recent works on video understanding and multimodal affective analysis have won various awards in international challenge evaluations, including CVPR ActivityNet Dense Video Captioning challenge, NIST TrecVID VTT evaluation, ACM Multimedia Audio-Visual Emotion Challenge etc.
Talk Title:Multimodal Emotion Recognition
Abstract:
Understanding human emotions is one of the fundamental steps in establishing natural human-computer interaction systems that
possess the emotion perception ability. The behavior signals of human emotion expression are multimodal, including voice,
facial expression, body language, bio-signals etc. and interactive scenarios such as conversational dialogues are the natural scenes
of emotion stimulation and expression. Our research focuses on the integration of multimodal information for robust emotion perception in
natural dialogues, which involves some of the major challenges, including modeling the contextual information in a dialogue, handling missing
modality issues in inference, learning robust emotion feature representation across cultures etc. This talk will present our recent works addressing
these challenges on multimodal emotion recognition.
Erik Cambria is the Founder of SenticNet, a Singapore-based company offering B2B sentiment analysis services, and an Associate Professor at NTU, where he also holds the appointment of Provost Chair in Computer Science and Engineering. Prior to joining NTU, he worked at Microsoft Research Asia (Beijing) and HP Labs India (Bangalore) and earned his PhD through a joint programme between the University of Stirling and MIT Media Lab. His research focuses on the ensemble application of symbolic and subsymbolic AI to natural language processing tasks such as sentiment analysis, dialogue systems, and financial forecasting. Erik is recipient of several awards, e.g., the 2019 IEEE Outstanding Early Career Award, he was listed among the 2018 AI's 10 to Watch, and was featured in Forbes as one of the 5 People Building Our AI Future. He is Associate Editor of many top AI journals, e.g., INFFUS, IEEE CIM, and KBS, Special Content Editor of FGCS, Department Editor of IEEE Intelligent Systems, and is involved in many international conferences as program chair and invited speaker.
Talk Title:Neurosymbolic AI for Affective Computing and Sentiment Analysis
Abstract:
Abstract: With the recent developments of deep learning, AI research has gained new vigor and prominence. However, machine learning still faces three big
challenges: (1) it requires a lot of training data and is domain-dependent; (2) different types of training or parameter tweaking leads to inconsistent
results; (3) the use of black-box algorithms makes the reasoning process uninterpretable. At SenticNet, we address such issues in the context of NLP via
sentic computing, a multidisciplinary approach that aims to bridge the gap between statistical NLP and the many other disciplines necessary for
understanding human language such as linguistics, commonsense reasoning, and affective computing. Sentic computing is both top-down and bottom-up:
top-down because it leverages symbolic models such as semantic networks and conceptual dependency representations to encode meaning; bottom-up
because it uses subsymbolic methods such as deep neural networks and multiple kernel learning to infer syntactic patterns from data.
Important Dates
Events | Dates |
---|---|
Submission of Manuscripts | August 31th, 2021 |
Notification of Acceptance/Rejection | September 9th, 2021 |
Submission for Camera-ready Papers and Presenting Author's Registration | September 15th, 2021 |
Date of the Workshop | October 18-22, 2021 |
Technical Program
-
PART 1: Opening
- 08:00 - 08:10
Opening statement
Dr. Youjun Xiong - 08:10 - 09:10
Title:Neurosymbolic AI for Affective Computing and Sentiment Analysis
Erik Cambria - 09:10 - 09:30
BERT Based Cross-Task Sentiment Analysis with Adversarial Learning
Zhiwei He, Xiangmin Xu, Xiaofen Xing and Yirong Chen - 09:30 - 09:50
Facial Micro-Expression Recognition Based on Multi-Scale Temporal and Spatial Features
Hao Zhang, Bin Liu, Jianhua Tao and Zhao Lv - 09:50 - 10:10
Aspect based sentiment analysis is a branch of sentiment analysis
Yingtao Huo, Dongmei Jiang and Hichem Sahli - 10:10 - 10:30
Call For Help Detection In Emergent Situations Using Keyword Spotting And Paralinguistic Analysis
Huangrui Chu, Yechen Wang, Ran Ju, Yan Jia, Haoxu Wang, Ming Li and Qi Deng - 10:30 - 11:30
Title: Multimodal Emotion Recognition
Qin Jin - 11:30 - 11:50
FER by Modeling the Conditional Independence between the Spatial Cues and the Spatial Attention Distributions
Wan Ding, Dongyan Huang, Jingjun Liang, Jinlong Jiao and Zhiping Zhao - 11:50 - 12:10
Efficient Gradient-based Neural Architecture Search for end-to-end ASR
Xian Shi, Pan Zhou, Wei Chen and Lei Xie - 12:10 - 12:30
Temporal Attentive Adversarial Domain Adaption for Cross Cultural Affect Recognition
Haifeng Chen, Yifan Deng and Dongmei Jiang - 12:30 - 12:50
A Multimodal Dynamic Neural Network for Call for Help Recognition in Elevators
Ran Ju, Huangrui Chu, Yechen Wang, Qi Deng, Ming Cheng and Ming Li - 12:50 - 13:10
A Web-Based Longitudinal Mental Health Monitoring System
Zhiwei Chen, Weizhao Yang, Jinrong Li, Jiale Wang, Shuai Li, Ziwen Wang and Lei Xie - 13:20 - 13:40
Semantic and Acoustic-Prosodic Entrainment of Dialogues in Service Scenarios
Liu Yuning, Jianwu Dang, Aijun Li and Di Zhou - 13:40 - 14:00
Improving Model Stability and Training Efficiency in Fast Speed High Quality Expressive Voice Conversion System
Zhiyuan zhao, jingjun liang, zehong zheng, linhuang yan, zhiyong yang, wan ding, and dongyan huang - 14:00 - 14:20
TeNC: Low Bit-Rate Speech Coding with VQ-VAE and GAN
Yi Chen, Shan Yang, Na Hu, Lei Xie and Dan Su - 14:20 - 14:40
Noise Robust Singing Voice Synthesis Using Gaussian Mixture Variational Autoencoder
Heyang Xue, Xiao Zhang, Jie Wu, Jian Luan, Yujun Wang and Lei Xie - 14:40 - 15:00
Dong-Yan Huang
Session Chair: Dong-Yan Huang
PART 2: Keynote Ⅰ
Session Chair: Björn Schuller
PART 3: Sentiment, Micro-expression and Paralinguistic analysis
Session Chair: Lei Xie
PART 4: Keynote Ⅱ
Session Chair: Dong-Yan Huang
PART 5: Multimodal Emotion Recognition
Session Chair: Jianhua Tao
PART 6: Multimodal Emotion Synthesis
Session Chair: Jie Yang
PART 7: Panel discussion and closing remarks
Submission Instructions
All the workshop accepted papers will be published in an adjunct proceedings to the 23nd ACM International Conference on Multimodal Interaction (ICMI 2021::23rd ACM International Conference on Multimodal Interaction). We invite the submissions in the following categories:
- Long papers/Short papers relevant to the themes of the workshop
- 1.5-pages* abstracts relevant to the themes of the workshop.
Format
Please note that ICMI 2021 will use a new ACM Publication System (TAPS) process.
This means submitted papers should follow the instructions and use templates given in the following link:
https://www.acm.org/publications/taps/word-template-workflow
All authors should submit manuscripts for review in a single-column format instead of the previous two-column format.
We invite submissions of long papers, short papers, and extended abstracts formatted according to the new
ICMI guidelines. It means:
- Long papers (8 pages in two-column -> 13~14 pages in one column + references)
- Short papers (4 pages in two-column) -> 7 pages in one column+ references)
- Extended abstracts (1 page in two-column) -> 1.5 pages in one column, Abstracts can contain Figures and/or Tables.
Template
Please refer to the instructions on the http://icmi.acm.org/2021/index.php?id=authors.
- (Word template) Write your paper using the Submission Template (https://www.acm.org/publications/taps/word-template-workflow). Follow the embedded instructions to apply the paragraph styles to your various text elements. The text is in single-column format at this stage and no additional formatting is required at this point.
- (Latex template) Please use the latest version of the Primary Article Template - LaTeX (1.78; published May 25, 2021) to create your paper submission. Use the "manuscript" call to create a single column format.
Publication
Proceedings will be published as adjunct proceedings to the ACM International Conference on Multimodal Interaction 2021.
Paper submission and acceptance deadlines
Workshop papers or abstract due: August 16th, 2021 (extended to August 31th, 2021)
Notification of acceptance: September 9th, 2021
Camera-ready paper: September 15th, 2021