Novel Framework for Multimodal Classification of Consumer Preference using EEG signals and eye-tracking data

Read the full article See related articles

Discuss this preprint

Start a discussion What are Sciety discussions?

Listed in

This article is not in any list yet, why not save it to one of your lists.
Log in to save this article

Abstract

Marketing plays a crucial role in enhancing consumer awareness and affinity towards a product. Consumer research examines the motivations, trends, and patterns underlying purchasing behavior, thereby enabling institutions to refine their offerings and marketing strategies to improve their alignment with consumer decision-making and increase purchase probability. Consumer neuroscience aims to study the neurological factors influencing consumer preference, which cannot be determined using traditional methods like questionnaires and self-assessment. Machine learning techniques have proven effective in this domain, yet deep learning architectures that effectively fuse electroencephalography (EEG) signals with eye-tracking (ET) data remain largely unexplored. In this study, we propose a novel hybrid "Wide & Deep" framework for multimodal classification of consumerpreference using EEG and eye-tracking data from the Neuma dataset. The proposed architecture employs an InceptionTime network with early fusion to process concatenated multimodal input (25 channels), extracting multi-scale temporal featuresthrough parallel convolutions with varying kernel sizes (10, 20, 40). The learned 128-dimensional neural embeddings are combined with 44 time-domain statistical features (temporal mean and standard deviation across all channels) and passed to a CatBoost gradient boosting classifier, forming a neural-symbolic ensemble. A modality dropout strategy (30% probability) is introduced during training to prevent over-reliance on eye-tracking signals and force robust EEG representation learning. The framework employs a grokking-oriented training regime with extended epochs (100), high weight decay (0.02), and strict checkpointing based on validation loss. The model is evaluated using 5-fold stratified cross-validation to assess generalizability. The proposed framework achieves a mean accuracy of 77.51% (±0.49%) and a weighted F1-score of 74.60% (±0.21%) in classifying consumer preferences, representing a significant improvement over prior methods on this dataset. This study demonstrates the effectiveness of multi-scale convolutional architectures combined with gradient boosting ensemble methodsfor multimodal consumer neuroscience applications.

Article activity feed