Abstract
Interpreting and managing social interactions is vital for social well-being, yet existing technologies fall short, particularly in group settings. This research aims to develop advanced machine perception systems for Social Signal Processing to accurately model human social behavior. Our multi-modal generative model aims to integrate multi-modal sensory data input data, contextual information and subjective observers’ narratives, utilizing them as complex input to an adapted Large Language Model, and producing plausible narratives that refect various human perspectives. This human-centered approach leverages both low-level cues and high-order events, ensuring adaptability to diverse observers and contexts. The model’s potential areas of application include cross-cultural interactions, social group integration, and professional meetings, enhancing social harmony and productivity.
Original language | English |
---|---|
Title of host publication | ICMI 2024 - Proceedings of the 26th International Conference on Multimodal Interaction |
Publisher | Association for Computing Machinery (ACM) |
Pages | 622-626 |
Number of pages | 5 |
ISBN (Electronic) | 9798400704628 |
DOIs | |
Publication status | Published - 2024 |
Event | 26th International Conference on Multimodal Interaction, ICMI 2024: ICMI 2024 - San Jose, Costa Rica Duration: 4 Nov 2024 → 8 Nov 2024 https://icmi.acm.org/2024/ |
Publication series
Name | ACM International Conference Proceeding Series |
---|
Conference
Conference | 26th International Conference on Multimodal Interaction, ICMI 2024 |
---|---|
Country/Territory | Costa Rica |
City | San Jose |
Period | 4/11/24 → 8/11/24 |
Internet address |
Keywords
- Human-centered computing
- Large Language Models
- Social Signal Processing