Big Data technology has been one of key engines driving the new industrial revolution. However, the majority of current Big Data research efforts have been devoted to single-modal data analysis, which leads to a huge gap in performance when algorithms are carried out separately. Although significant progress has been made, single-modal data is often insufficient to derive accurate and robust models in many applications.

Multimodal is the most general form for information representation and delivery in a real world. Using multimodal data is natural for humans to make accurate perceptions and decisions. In fact, our digital world is multimodal, combining different modalities of data such as text, audio, images, videos, animations, drawings, depth, 3D, biometrics, interactive content, etc. Multimodal data analytics algorithms often outperform single modal data analytics in many real-world problems.

Multi-sensor information fusion has also been a topic of great interest in industry nowadays. In particular, such companies working on automotive, drone vision, surveillance or robotics have grown exponentially. They are attempting to automate processes by using a wide variety of control signals from various sources.

With the rapid development of Big Data technology and its remarkable applications to many fields, multimodal Big Data is a timely topic. This workshop aims to generate momentum around this topic of growing interest, and to encourage interdisciplinary interaction and collaboration between Natural Language Processing (NLP), computer vision, audio processing, machine learning, multimedia, robotics, Human-Computer Interaction (HCI), cloud computing, Internet of Things (IoT), and geospatial computing communities. It serves as a forum to bring together active researchers and practitioners from academia and industry to share their recent advances in this promising area.

Topics

This is an open call for papers, which solicits original contributions considering recent findings in theory, methodologies, and applications in the field of multimodal Big Data. The list of topics includes, but not limited to:

  • Multimodal data modeling
  • Multimodal learning
  • Cross-modal learning
  • Multimodal big data analytics
  • Multimodal big data infrastructure and management
  • Multimodal scene understanding
  • Multimodal data fusion and data representation
  • Multimodal perception and interaction
  • Multi-modal benchmark datasets and evaluations
  • Multimodal information tracking, retrieval and identification
  • Multimodal object detection, classification, recognition and segmentation
  • Language and vision (e.g., image/video searching and captioning, visual question answering, visual scene understanding, etc.)
  • Language and audio
  • Audio and vision
  • Biometrics and big data (e.g., face recognition, behavior recognition, eye retina and movement, palm vein and print, etc.)
  • Multimodal applications (e.g., autonomous driving, robotic vision, cybersecurity, smart cities, industrial inspection, medical diagnosis, social media, arts, etc.)


Important Dates

Early submission and notification dates (Anywhere on Earth):
  • Oct. 2, 2023: Submission of full papers (9-10 pages)
  • Oct. 9, 2023: Submission of short papers (6 pages + 1- or 2-page references & appendices)
  • Oct. 16, 2023: Submission of poster papers (3-4 pages) or extended abstracts (5 pages)
  • Nov 1, 2023: Notification of paper acceptance
  • Nov. 10, 2023: Submission of revisions of conditionally accepted papers/posters for the second round of review
Regular submission and notification dates (Anywhere on Earth):
  • Oct. 16, 2023: Submission of full papers (9-10 pages)
  • Oct. 23, 2023: short papers (6 pages + 1- or 2-page references & appendices)
  • Oct. 30, 2023: Submission of poster papers (3-4 pages) or extended abstracts (5 pages)
  • Nov 12, 2023: Notification of paper acceptance
  • Nov. 19, 2023: Submission of revisions of conditionally accepted papers/posters for the second round of review
  • Please check accepted papers/posters. The acceptance notifications for the second-round review were sent out via email. Please incorporate review comments into your revisions and submit them by the deadline specified in your email for final review before camera-ready submission.

Final camera-ready submission dates (Anywhere on Earth):


Submission

Please follow IEEE manuscript templates (Overleaf or US Letter) and IEEE reference guide to format your paper, and then directly submit to IEEE Big Data paper submission site.
The submissions must be in PDF format without author list (double-blind), written in English, and formatted according to the IEEE publication camera-ready style. All the paper review follows double-blind peer review.
Accepted papers will be published in the IEEE Big Data 2023 proceedings.


Program Committee

Program Chair

Lindi Liao, George Mason University, USA

Program Committee Members

  • Zhiqian Chen, Mississippi State University, USA
  • Kaiqun Fu, South Dakota State University, USA
  • Jesse Guessford, George Mason University, USA
  • Maryam Heidari, George Mason University, USA
  • Fanchun Jin, Google Inc., USA
  • Ge Jin, Purdue University, USA
  • Ashwin Kannan, Amazon, USA
  • Kevin Lybarger, George Mason University, USA
  • Emanuela Marasco, George Mason University, USA
  • Nora McDonald, George Mason University, USA
  • Abhimanyu Mukerji, Amazon, USA
  • Chen Shen, Google Inc., USA
  • Arpit Sood, Meta, USA
  • Gregory Joseph Stein, George Mason University, USA
  • Marcos Zampieri, George Mason University, USA
  • Yanjia Zhang, Boston University, USA


Accepted Papers/Posters

Acceptance rate: 25% (full paper), 33% (short paper)

Early Submission

  • Multimodal Large Language Models: A Survey
  • PRE: Vision-Language Prompt Learning with Reparameterization Encoder
  • Understanding the Language of ADHD and Autism Communities on Social Media
  • Gender Classification Accuracy via Two-Dimensional Body Joints using Convolutional Neural Networks
  • Audio-visual Neural Face Generation with Emotional Stimuli
  • LoRA-like Calibration for Multimodal Deception Detection using ATSFace Data
  • Enhancing Scientific Image Classification through Multimodal Learning: Insights from Chest X-Ray and Atomic Force Microscopy Datasets
  • A Novel Time Series Anomaly Detection Method Based on Prototype Error
  • CLIP-PubOp: A CLIP-based Multimodal Representation Fusion Method for Public Opinion
  • Deep Explainability: Spin-Geometrical Neural Meta-Structures
  • De-SaTE: Denoising Self-attention Transformer Encoders
  • Stratified Stability Selection for Multi-Omics

Regular Submission

  • Evaluating CLIP's Understanding on Relationships in a Blocks World
  • Exploiting Multiple Sequence Lengths in Fast End to End Training for Image Captioning
  • Impact of Mixed Multimodalities and Size Dependence on Performance of Object Detection on Multimodal Satellite Imagery
  • New Spoof Finger Photo Data with Demographics for Enhanced Smartphone Security
  • Character-based Outfit Generation with Vision-augmented style extraction via LLMs
  • Predicting Potential School Shooters from Social Media Posts
  • Semantic Prompt Based Multi-Scale Transformer for Few-Shot Classification
  • Using the CARLA Simulator to Train A Deep Q Self-Driving Car to Control a Real-World Counterpart on A College Campus
  • Multimodal One-class Learning for Malicious Online Content Detection
  • A Supervised Autoencoder for Human Activity Recognition with Inertial Sensors
  • Late Fusion-based Distributed Multimodal Learning