Analysis of Deep Learning Models for EEG Data Processing

Slide Note
Embed
Share

This content delves into the application of deep learning models, such as Sequential Modeler, Feature Extraction, and Discriminator, for processing EEG data from the TUH EEG Corpus. The architecture involves various layers like Convolution, Max Pooling, ReLU activation, and Dropout. It explores temporal and spatial context analysis, feature extraction techniques, and classification results in the context of EEG data processing. The models are designed to handle long-term dependencies and generate predictions based on input EEG signals.


Uploaded on Sep 21, 2024 | 0 Views


Download Presentation

Please find below an Image/Link to download the presentation.

The content on the website is provided AS IS for your information and personal use only. It may not be sold, licensed, or shared on other websites without obtaining consent from the author. Download presentation by click this link. If you encounter any issues during the download, it is possible that the publisher has removed the file from their server.

E N D

Presentation Transcript


  1. Blind Evaluation Blind Evaluation Feature Extraction Select Model Blind Evaluation Feature Extraction Select Model Collect Data Train Feature Extraction Select Model Collect Data Validate Train Collect Data Train Validate Validate

  2. Epoch Posteriors Temporal and Spatial Context Analysis by Deep Learning ? Sequential Modeler ?11 ?22 ?33 ? ?? Feature Extraction ?12 ?23 TUH EEG ?? ??(x,z) ?2(?) ?1(?) ?3(?) Corpus ? ? Hidden Markov Models Stacked Denoising Autoencoders Post Processor? ? Output S2 S1 ? ? ? ? S4 S3 ? Epoch Label Finite State Machine for Language Model

  3. Input Layer Convolution Max Pooling 16@34*10 Convolution Convolution Convolution 32@32*8 16@68*20 32@34*10 16@70*22 Channels Flatten Max Pooling Convolution Max Pooling 32@16*4 Convolution Time 64@16*4 64@14*2 64@7*1 Output ...

  4. Input ReLU ReLU ReLU Conv2d Dropout Conv2d Dropout Conv2d Dropout Conv2d ReLU ReLU Dropout Conv2d ReLU Dropout Conv2d ReLU Dropout Conv2d Conv2d ReLU Dropout Conv2d + + + + Max Pooling2D Max Pooling2D ReLU ReLU Dropout Conv2d Dropout Conv2d Flatten ReLU ReLU ReLU Dropout Conv2d ReLU Dropout Conv2d Dense Dropout Conv2d Dropout Sigmoid + + Predictions

  5. Discriminator Classification Result CNN Real World EEGs TUH EEG Corpus x Transposed CNN Is Discriminator Correct? Noise Synthetic Generated EEGs Update Models Generator

  6. Sequential Modeler Feature Extraction Output Gate ?? Output Input Gate ?? ?? ?? TUH EEG Corpus Forget Gate ?? Learning Long-Term Dependencies with LSTM

  7. Feature Extraction 2D-Max Pooling 210@5*6*32 2D-Convolution 2D-Max Pooling 210@11*13*16 2D-Convolution 210@22*26*16 210@11*13*32 210@22*26*1 TUH EEG Corpus 2D-Convolution 1D-Convolution 210@16 Flatten 210@384 2D-Max Pooling 210@2*3*64 210@5*6*64 Sequential Modeler Post Output Gate Processor ?? Input Gate ?? Output ?? ?? 1D-Max Pooling 26@16 Forget Gate ?? Learning Long-Term Dependencies with B- LSTM

  8. Epoch Posteriors Sequential Modeler Output Gate Sequential Modeler ?11 ?? ?22 ?33 Input Gate ?? Feature Extraction ?12 ?23 TUH EEG ?? ?? ?2(?) ?1(?) ?3(?) Corpus Forget Gate Hidden Markov Models ?? Learning Long-Term Dependencies with LSTM Post Processor? ? Output S2 S1 ? ? ? ? S4 S3 ? Epoch Label Finite State Machine for Language Model

  9. 0.88 0.9 7 0.61 0.82 0.91 0.87 0.72 0.57 0.77 0.99 0.6 Probabilistic filtering (Threshold = 0.75) 0.88 0.9 7 0.82 0.91 0.87 0.77 0.66 5 0.99 0.6 Duration filtering (Threshold = 3 Sec.) 0.88 0.63 2 0.9 7 0.82 0.91 0.87 0.77 Target aggregation (Margin = 3.54 Sec.) 0.88 0.63 2 0.9 7 0.91 0.82 Posterio r Seizure Posterio r Backgroun d Time scale 4 Seconds

Related