No CrossRef data available.
Published online by Cambridge University Press: 05 January 2022
Background: Intracranial electroencephalography (iEEG) recordings are obtained from the sampling of sub-cortical structures and provide extraordinary insight into the spatiotemporal dynamics of the brain. As these recordings are increasingly obtained at higher channel counts and greater sampling frequencies, preprocessing through visual inspection is becoming untenable. Consequently, artificial neural networks (ANNs) are now being leveraged for this task. Methods: One-hour recordings from six patients diagnosed with drug-resistant epilepsy at Toronto Western Hospital were obtained alongside fiduciary ECG and EOG activity. R-wave peaks and local maxima were identified in the ECG and EOG recordings, respectively, and were time-mapped onto the iEEG recordings to delimit one-second epochs around 1.6 million cardiac and 600 thousand ocular artifacts. Epochs were then split into train-test-evaluation sets and fed into an ANN as one-second spectrograms (0 - 1,000 Hz) over 30-time steps. Results: The ANN model achieved formidable classification results on the evaluation set with an F1, positive predictive value, and sensitivity scores of 0.93. Furthermore, model architecture computed the classification probability at each time-step and enabled insight into the spatiotemporal features driving classification. Conclusions: We expect this research to promote the public sharing of new ANN from multiple institutions and enable novel automated algorithms for artifact detection in iEEG recordings.