TY - JOUR
T1 - Crowdsourcing reproducible seizure forecasting in human and canine epilepsy
AU - Brinkmann, Benjamin H.
AU - Wagenaar, Joost
AU - Abbot, Drew
AU - Adkins, Phillip
AU - Bosshard, Simone C.
AU - Chen, Min
AU - Tieng, Quang M.
AU - He, Jialune
AU - Muñoz-Almaraz, F. J.
AU - Botella-Rocamora, Paloma
AU - Pardo, Juan
AU - Zamora-Martinez, Francisco
AU - Hills, Michael
AU - Wu, Wei
AU - Korshunova, Iryna
AU - Cukierski, Will
AU - Vite, Charles
AU - Patterson, Edward E.
AU - Litt, Brian
AU - Worrell, Gregory A.
N1 - Funding Information:
The authors acknowledge the generous support of the American Epilepsy Society, The Epilepsy Foundation, Kaggle.com (which waived a portion of its normal fee for this competition), and the National Institutes of Health. Data collection, processing, analysis, and manuscript preparation were supported by NeuroVista Inc. and grants NIH-NINDS UH2/UH3 95495 (G.W.), U01-NS 73557 (G.W.), U24-NS063930 (B.L., G.W.), K01 ES025436-01 (J.W.), and R01-NS92882 (G.W.), the Mirowski family foundation, and Mayo Clinic.
PY - 2016/6
Y1 - 2016/6
N2 - Accurate forecasting of epileptic seizures has the potential to transform clinical epilepsy care. However, progress toward reliable seizure forecasting has been hampered by lack of open access to long duration recordings with an adequate number of seizures for investigators to rigorously compare algorithms and results. A seizure forecasting competition was conducted on kaggle.com using open access chronic ambulatory intracranial electroencephalography from five canines with naturally occurring epilepsy and two humans undergoing prolonged wide bandwidth intracranial electroencephalographic monitoring. Data were provided to participants as 10-min interictal and preictal clips, with approximately half of the 60GB data bundle labelled (interictal/preictal) for algorithm training and half unlabelled for evaluation. The contestants developed custom algorithms and uploaded their classifications (interictal/preictal) for the unknown testing data, and a randomly selected 40% of data segments were scored and results broadcasted on a public leader board. The contest ran from August to November 2014, and 654 participants submitted 17 856 classifications of the unlabelled test data. The top performing entry scored 0.84 area under the classification curve. Following the contest, additional held-out unlabelled data clips were provided to the top 10 participants and they submitted classifications for the new unseen data. The resulting area under the classification curves were well above chance forecasting, but did show a mean 6.54 2.45% (min, max: 0.30, 20.2) decline in performance. The kaggle.com model using open access data and algorithms generated reproducible research that advanced seizure forecasting. The overall performance from multiple contestants on unseen data was better than a random predictor, and demonstrates the feasibility of seizure forecasting in canine and human epilepsy.
AB - Accurate forecasting of epileptic seizures has the potential to transform clinical epilepsy care. However, progress toward reliable seizure forecasting has been hampered by lack of open access to long duration recordings with an adequate number of seizures for investigators to rigorously compare algorithms and results. A seizure forecasting competition was conducted on kaggle.com using open access chronic ambulatory intracranial electroencephalography from five canines with naturally occurring epilepsy and two humans undergoing prolonged wide bandwidth intracranial electroencephalographic monitoring. Data were provided to participants as 10-min interictal and preictal clips, with approximately half of the 60GB data bundle labelled (interictal/preictal) for algorithm training and half unlabelled for evaluation. The contestants developed custom algorithms and uploaded their classifications (interictal/preictal) for the unknown testing data, and a randomly selected 40% of data segments were scored and results broadcasted on a public leader board. The contest ran from August to November 2014, and 654 participants submitted 17 856 classifications of the unlabelled test data. The top performing entry scored 0.84 area under the classification curve. Following the contest, additional held-out unlabelled data clips were provided to the top 10 participants and they submitted classifications for the new unseen data. The resulting area under the classification curves were well above chance forecasting, but did show a mean 6.54 2.45% (min, max: 0.30, 20.2) decline in performance. The kaggle.com model using open access data and algorithms generated reproducible research that advanced seizure forecasting. The overall performance from multiple contestants on unseen data was better than a random predictor, and demonstrates the feasibility of seizure forecasting in canine and human epilepsy.
KW - Epilepsy
KW - Experimental models
KW - Intracranial EEG
KW - Refractory epilepsy
UR - http://www.scopus.com/inward/record.url?scp=84978888933&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84978888933&partnerID=8YFLogxK
U2 - 10.1093/brain/aww045
DO - 10.1093/brain/aww045
M3 - Article
C2 - 27034258
AN - SCOPUS:84978888933
SN - 0006-8950
VL - 139
SP - 1713
EP - 1722
JO - Brain
JF - Brain
IS - 6
ER -