This repository is inspired on the original implementation of SincNet. Here we modified and adapted the original SincNet code to evaluate the performance of a SincNet-based architecture for EEG-based emotion recognition. For this evaluation we utilized EEG data collected on the Social Competence and Treatment Lab (SCTL) from StonyBrook University, NY, USA. The EEG data were collected from 40 individual diagnosed with Autism-Spectrum-Disorder (ASD) and 48 typically-developed (TD) or non-ASD participants. For any data request (e.g., behavioral, EEG) or analysis, pleased send an email to the corresponding contributors at [email protected], [email protected], or [email protected] and be patient during the request. The original SincNet paper describing the architecture and the corresponding test on speech and speaker recognition is here Ravanelli & Bengio 2018.
These is are the mandatory requirements for this code repository:
- Python >= 3.6
- Pytorch >= 1.0
Look both packages are compatible before doing the installation and avoid unnecessary forced errors.
If you want to use this pipeline and/or request the dataset associated to this repository please cite our paper Mayor-Torres, J. M., Ravanelli, M., Medina-DeVilliers, S. E., Lerner, M. D., & Riccardi, G. (2021). Interpretable SincNet-based Deep Learning for Emotion Recognition from EEG brain activity. 2021 43rd Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC)
The proposed SincNet-based architecture for EEG-based emotion recognition is described on the following Figure:
This system is composed of a SincConv layer, three standard conv-pool blocks such as Conv1, Conv2, and Conv3, and a fully-connected DNN1 coupled and connected to a softmax classifier. The three 2D convolutional blocks are composed of 32, 64, and 128 channels with kernel sizes of (100 x 10), (20 x 5), (5 x 2), respectively. Max-pooling used kernel sizes of (10 x 5), (5 x 2), (2 x 2) are also included in the pipeline. All the DNN units were ReLUs.
After you downloaded the data from the request you can install the package requirements on the SincNet repository and run the leave-one-trial-out (LOTO) cross-validation using the following command:
python main_SincNet_EEG_model.py "/folder/with_TD_or_ASD_ZCA_inputs/"
Please be careful where you set up your data folder containing the 40, 48 or any amount of participant data you want to use in your experiment. This process can take a lot of execution time and memory. For a P100 GPU the full simulation on the 88 participants took approximately four complete days. Running the process the results will be saved on temporary files with the name "res_evaluation_subjectID_trialnumber.csv" containing the loss and the accuracy average metrics, and files with the name "res_error_subjectID_trialnumber.csv" contains the binary representation of hits when the test trial label matches with the prediction and a miss otherwise.
The performance comparison between the human (participant) performance, a CNN baseline described here, and the SincNet-based system is plotted using a barplot in the following Figure:
An important feature of our proposed SincNet-based system is to learn the frequency responses from TD and ASD participants using a by-design interpretability resource. This consists in an application of a Fast-Fourier-Transform on each filters' impulse-response learned from the SincConv layer. The following .gif loop shows the Power-Spectral-Density (PSD) for TD (blue) and ASD (red) individuals across all the 400 training loops we used on this experiment. This information is learned in unsupervised (implied) way without adding any diagnosis label.
To evaluate the SincLayer filter weights after training, we attached them for each 400 iteration (training loops) and for each trial per participant 1-48. You can calculate the PSD from the read_filters_act_more.m code. These SincLayer's weights and its corresponding files are located in the following two Google Drive links. These files are developed with an experiment implement with a size of the filters in the SincLayer of 50. You can calculate your own filters using the line 426 of the main_EEG_SincNet.py code.
- https://drive.google.com/file/d/1oRFI0lGdqxDOMsLhNyO8bLSfVB8-U4E-/view?usp=sharing
- https://drive.google.com/file/d/1Xn1xJuYDOJnlh5YrB8jSt05p9wsVfrAi/view?usp=sharing
After 400 training iterations we obtained significant differences on high-α (9-13 Hz) F(1,87)=3.331,p=0.000267 and β (13-30 Hz) F(1,87)=2.05,p=0.00102 bands after Bonferroni-Holm correction. This significant difference are consistent with previous EEG studies which includes data from individuals with ASD Pineda et., al 2012, Friederich et., al 2015, and Levin et., al 2020.
To process all the results for filter activation, metrics and results wrapping use all the code released in the m_files directory. To find the statistical difference between the frequency responses you need to run the read_filters_act_more.m function for each subject in the dataset, and subsequently you need to run the evaluate_differences_filters.m function, before you identify the folders you saved the filters files for TD and ASD groups.