英语轻松读发新版了,欢迎下载、更新

A systematic review of EEG-based machine learning classifications for obsessive-compulsive disorder: current status and future directions

2025-09-09 11:33:41 英文原文

作者:Jahanian-Najafabadi, Amir

  • Systematic Review
  • Open access
  • Published:

BMC Psychiatry volume 25, Article number: 854 (2025) Cite this article

Abstract

Obsessive–compulsive disorder (OCD) is a chronic and disabling condition affecting approximately 3.5% of the global population, with diagnosis on average delayed by 7.1 years or often confounded with other psychiatric disorders. Advances in electroencephalography (EEG) analysis using machine learning hold promise for the development of OCD-specific biological markers. This systematic review aims to evaluate studies that classify individuals with OCD from other groups based on EEG data. Following PRISMA guidelines, we searched the Web of Science, Scopus, PubMed, and IEEE databases through February 2025; of 42 screened studies, 11 met inclusion criteria for final analysis. Data were extracted across four domains: general information, population characteristics, EEG features, and machine learning features. Results revealed extensive heterogeneity in study populations, associated symptoms, EEG preprocessing methods, validation strategies, and reporting of model accuracy, underscoring the need for harmonized standards. Notably, only a few studies provided statistical interpretation of their models. None of reviewed studies employed modern interpretability techniques such as SHAP or LIME methods that, beyond reducing “black-box” opacity, can inform optimal electrode placement for neurofeedback or transcranial electrical stimulation. Many studies were constrained by cultural limitations, small sample sizes and lack of demographic information e.g., age, gender, medication. This work represents the first systematic review of EEG-ML classification studies in OCD and emphasizes the urgent need for methodological standardization in this emerging field.

Peer Review reports

Introduction

Obsessive-Compulsive Disorder (OCD) is a common and long-lasting mental health condition that affects a person’s social life, family, and society [1]. About 60% of people with OCD experience chronic symptoms in their lives [2]. OCD affects around 3.5% of people, with women (5.4%) more likely to have it than men (1.7%) [2]. A study found that half of OCD symptoms start by age 18, and 70% start before age 20. This shows why looking for early signs in teenagers is important [2]. OCD greatly reduces a person’s quality of life and work performance [2].

More than half of patients diagnosed with OCD have serious problems in family relationships and daily activities because of the symptoms [3]. The costs of OCD both direct e.g., cost of treatment and indirect such as lost work and associated income are very high [3]. Because of this, the World Health Organization (WHO) listed OCD as one of the most disabling mental disorders which extremely affect life of individuals who suffer from OCD sysmptoms and their families [4]. Even though OCD is serious psychiatric condition, many people do not get diagnosed or treated early in life. On average, it takes 7.1 years from the onset of symptoms to the diagnosis stage [5]. Reasons for this delay include social stigma, limited healthcare access, and financial problems [6]. Additionally, OCD can be misdiagnosed by specialists due to the similarity and overlaps of symptoms with other psychiatric disorders, or potentially due to the feeling of shame in patients [7]. A study found that 33.8% of African-American patients with OCD were not diagnosed correctly and with precision. This may be because common tools for diagnosis do not work well for this group [8].

For treatment-resistant OCD, methods like antipsychotic drugs, deep brain stimulation, and neurosurgical interventions can help [9]showing the importance of brain-focused treatments. EEG with high temporal resolution at the millisecond scale is beneficial because it can provide biological markers and show how the brain is affected both anatomically and functionally [10]. Studies using EEG have found irregularities in the frontotemporal brain areas, especially in women and those who respond to treatment [11]. EEG is also affordable [12, 13]easy to use, and helpful in emergency conditions [13]. Application of machine learning techniques in classification of OCD compared to healthy individuals and to distinguish with other psychiatric conditions suggests a new avenue to improve the diagnostic tools for mental disorders e.g., OCD.

With advanced computational modelings and machine learning techniques we are able to detect unusual/abnormal brain patterns and help diagnose mental disorders [14, 15]. These methods can also suggest personalized treatments based on brain activity [16, 17]. However, challenges like data standardization and model accuracy still exist [15, 16, 18]. There have been several computational studies in recent years that have aimed to either distinguish healthy from other neuropsychiatric and neurodevelopmental disorders, e.g., ADHD [19]; Autism Spectrum Disorder [20, 21]; MDD [22,23,24] or for time-series forecasting in order to predict EEG signals [25].

In order to apply machine learning models in practice, their interpretability is of great importance. Modern interpretability techniques such as SHAP (SHapley Additive exPlanations) and LIME (Local Interpretable Model-agnostic Explanations), by identifying the relative importance of features in the model’s prediction, can provide researchers and clinicians with a better understanding of the contribution of each variable for instance, the relative importance of EEG electrodes within different frequency bands. LIME generates synthetic samples around the target instance and fits a simple local linear model to the outputs of the main model to extract the local influence of each feature, while SHAP calculates the fair contribution of each feature across all possible feature combinations, assigning importance values that are disentangled and independent from other features [26,27,28]. For example, recent studies in the fields of EEG in response to treatment [29]classification of stroke [30]and diagnosis of depression [31] have effectively utilized SHAP and LIME to analyze and determine the importance of EEG features. However, to our knowledge, this is the first systematic review focused on EEG-based machine learning for classification of OCD compared to healthy(other) group(s). In this research, we examined recent progress and highlights the need for improvement of research methods and filling knowledge gaps.

Methods

Search strategy

Following PRISMA guidelines, a systematic search was conducted in PubMed, Scopus, IEEE, and Web of Science databases from its research inception to 6 February 2025. After the search, 67 articles were found (cf., Fig. 1). These articles were imported into EndNote software and duplicates were removed. After removing duplicates, 42 articles remained. These 42 articles were extracted and imported into Rayyan and we reviewed independently to include or exclude articles based on the inclusion/exclusion criteria. If there was a disagreement, we further evaluated and discussed until we agreed. Finally, 11 articles were included in this study. The accepted articles were entered into an Excel file with four main areas:

Fig. 1
figure 1

PRISMA flowchart for the selection of studies

  • General information (Table 1).

    Table 1 General information of included studies
  • Population (Table 2).

    Table 2 Population characteristics and clinical demographics
  • EEG features (Table 3).

    Table 3 EEG recording protocols and feature extraction methods
  • Machine learning features (Table 4).

    Table 4 Machine learning methodologies and performance metrics

In the next step, the extracted data were interpreted qualitatively. A quantitative meta-analysis was not performed due to substantial heterogeneity across studies. Specifically, the sample sizes, and that the reviewed studies employed a wide variety of EEG preprocessing pipelines, including different filtering, artifact removal, and feature extraction techniques. In addition, the machine learning models varied considerably, encompassing SVMs, Random Forests, Convolutional Neural Networks (CNNs), and ensemble methods, with diverse validation strategies such as 10-fold cross-validation, leave-one-out, and hold-out methods. These methodological differences, along with inconsistent reporting of effect sizes and outcomes, precluded meaningful statistical synthesis. While no formal heterogeneity statistics could be calculated due to insufficient comparable data, the degree of variation in design and analysis across studies supports the decision to pursue a qualitative synthesis.

Search terms

(“EEG” OR “electroencephalography” OR “resting state EEG” OR “resting-state EEG”) AND (“machine learning” OR “artificial intelligence” OR “deep learning”) AND (“Obsessive compulsive disorder” OR “OCD” OR “Obsessive-Compulsive”) AND (“classification” OR “diagnosis” OR “pattern recognition” OR “taxonomy” OR “Convolutional Neural Network” OR “Random Forest” OR “Short Long Term Memory” OR “LSTM” OR “Computational Modeling” OR “Early diagnosis”).

Inclusion/Exclusion criteria

Inclusion criteria:

  • Studies with OCD population.

  • Use of machine learning techniques applied to OCD-related EEG data.

  • Peer-reviewed.

  • Full-text available.

  • Published in English.

  • Publication date from database inception up to February 2025.

Exclusion criteria:

  • Studies not directly applying machine learning for classification.

  • Studies not focusing on OCD patients.

  • Studies not using EEG for OCD-related classification.

  • Non-English articles.

  • Articles that are not original research.

  • Case reports.

Quality of studies

To check the risk of bias in the studies we included, we used a changed version of the QUADAS-2 tool. This tool was first made for medical diagnosis studies, but we used it for three main reasons. First, we could change the questions to match ML-EEG studies, but still keep the four main parts: how patients were chosen, how the test was done, how OCD was diagnosed, and how data was used over time. Second, these parts match the important steps in EEG and machine learning research. Third, there is no special tool made for EEG-ML studies in psychiatry yet. So, QUADAS-2 was a good and flexible option. Two reviewers checked the bias separately, and they talked to agree if there were any differences. Table 5 present QUADAS-2 criteria employd for EEG-ML studies.

Table 5 Adapted QUADAS-2 criteria for EEG-ML studies

Risk of bias

The highest risk of bias was identified in the patient selection domain, while the lowest risk was observed in the flow and timing domain. Many studies appeared to prioritize an engineering-focused approach, emphasizing model development over detailed descriptions of patient populations. Several studies did not clearly report how OCD patients were screened. Random data splitting was confirmed in only five studies, while the remaining six did not provide sufficient information (cf., Table 6). Studies reporting very high accuracy such as Mukherjee et al. (2024) with 100% accuracy which also exhibited the highest risks of unclear bias. Conversely, studies with lower bias levels such as Erguzel et al. (2015) reporting 57–81% accuracy tended to align more closely with clinical reality. These findings suggest that minimizing bias should take precedence over simply maximizing accuracy.

Table 6 Risk of bias assessment using modified QUADAS-2 criteria

Result

In this systematic review, we examined studies that use EEG and machine learning to classify patients with OCD from other clinical conditions. As this is a relatively new emerging field, this review serves as a foundational step in identifying current gaps and offering suggestions for future EEG-ML research and reporting. Narrative synthesis of studies is shown in four areas:

  • General study information (Table 1).

  • Test population (Table 2).

  • EEG features (Table 3).

  • Machine learning features (Table 4).

This study follows PRISMA diagram to improve quality and transparency.

Emerging field

The application of EEG and machine learning in psychiatric diagnosis, particularly for OCD is a rapidly growing field. Prior to 2020, only one relevant study had been published. However, in the past five years, the number has increased to 11 studies, highlighting a significant rise in research and potentially clinical interest. This growth is largely driven by recent advancements in deep learning and machine learning techniques and better availability of datasets [32,33,34]. Despite this progress, substantial research gaps remain, underscoring the need for methodological and clinical refinement. The relatively small number of studies also emphasizes the nascent stage of this research domain.

Geographical spread and generalization issues

Among the 11 studies analyzed, 10 originated from Asian countries, with only one involving a Europe–Asia collaboration (cf., Fig. 2). This limited geographic distribution raises concerns about the generalizability of findings to non-Asian or Western populations. However, the use of publicly available datasets may mitigate some of this bias. Of the included studies, 36% (n = 4) used public datasets, 54% (n = 6) relied on original data collection, and 9% (n = 1) employed a combination of public and private data. The growing use of shared datasets represents a promising avenue for enhancing reproducibility and cross-cultural applicability in future research.

Fig. 2
figure 2

Visual Summary of Bias Risk Distribution

Different definitions of OCD

Approximately 54% of the reviewed studies did not report how OCD patients were screened, resulting in an “unclear” risk of bias rating for the reference standard domain in the QUADAS-2 assessment (cf., Table 6). Only 18% (n = 2) explicitly stated that psychiatrists conducted the screening, while 9% (n = 1) relied on self-report questionnaires. Of the six studies that omitted screening details, four used public datasets and two used privately collected data. Although public datasets improve accessibility and facilitate research, they often lack critical demographic and diagnostic information, which may introduce bias. These findings highlight the need for standardized, publicly available EEG datasets with clear documentation and clinical validation such as Deutsche Forschungsgemeinschaft (DFG)- based guidelines.

We identified a significant research gap in the reviewed studies which nearly all studies have overlooked DSM-4/DSM-5 OCD subtypes (such as contamination and symmetry) in EEG-based machine learning analyses. Treating OCD as a homogeneous disorder risks oversimplifying its neural underpinnings. Functional MRI research has demonstrated distinct brain activation patterns across different OCD subgroups [35, 36]. If similar subtype-specific patterns can be detected using EEG, it could pave the way for developing personalized neurofeedback and TES protocols tailored to individual symptom profiles.

None of the reviewed articles addressed OCD subtypes, such as symmetry or contamination obsessions, representing a notable research gap. Future studies should investigate whether different OCD subtypes exhibit distinct EEG patterns. Only one study categorized OCD patients based on high and low anxiety levels, but this classification did not align with DSM-defined subtypes. Seven studies (63%) reported details about the specific OCD populations included. Three studies (27%) did not mention OCD subtypes, and one study omitted demographic information entirely. Studies with smaller sample sizes. For example, Mukherjee et al. with 46 patients often reported unusually high classification accuracy (up to 100%), while those with larger samples such as Farhad et al. (2024). with 86 patients tended to report more moderate and stable accuracy rates (ranging from 75 to 90%).

EEG electrode montage and frequency bands involved

Eight studies used the standard 10–20 EEG montage, while two employed the higher-resolution 10–10 montage. One study did not specify the montage used. A significant research gap was the lack of information on reference electrodes: six studies (~ 54%) did not report which reference was used. This omission hinders reproducibility and limits the ability of other researchers to replicate or build upon these findings.

The classic four EEG bands (delta, theta, alpha, and beta) were analyzed in 54% of the studies. Three studies (27%) focused exclusively on these bands, while another three (27%) also included high beta and gamma frequencies. Two studies utilized raw EEG signals, and two did not specify which frequency bands were used. One study employed EEG microstate analysis. Notably, both studies that used raw signals applied deep learning techniques and required larger sample sizes, reflecting the increased data demands of such approaches.

Most studies focused on the resting-state condition, particularly with eyes closed. Approximately 81% (n = 9) of the articles used resting-state EEG; of these, seven studies employed the eyes-closed condition. One study combined resting-state and task-based recordings, while another did not specify the condition used. The eyes-closed resting-state may help minimize visual processing biases; however, it is also associated with increased alpha activity, which could influence the interpretation of EEG findings in OCD research.

Data preprocessing

Bandpass filtering and Independent Component Analysis (ICA) were among the most commonly used preprocessing techniques. Of the 11 studies, five applied bandpass filtering, and three used ICA. Reported classification accuracies in ICA-based studies ranged from 68 to 91%. However, due to variability in machine learning models and classification objectives, it is difficult to isolate the specific impact of ICA on performance. Two studies did not report any preprocessing steps. While proper preprocessing is generally believed to enhance machine learning accuracy, the inconsistency in methods across studies prevents firm conclusions. The adoption of standardized preprocessing guidelines, such as those proposed by COBIDAS-EEG, is recommended to improve comparability and reproducibility.

As shown in Table 7, earlier studies tend to report lower classification accuracies, while more recent publications show higher performance. One likely explanation is the ongoing development of machine learning techniques within psychiatric research. Over time, the field has moved from using relatively simple models to employing more advanced methods, such as deep learning and combinations of multiple algorithms. Alongside these advancements, there have also been notable improvements in EEG data processing. For example, in how noise and artifacts are removed, interpolation of channels in clinical data, and how relevant features are extracted from the signals.

Table 7 Summary of reported classification accuracies for OCD disorders using EEG-Based machine learning models

Another factor that may contribute to the improved results is the availability of more powerful computing tools and, in some cases, larger and better-quality datasets. These conditions make it easier to train and test models more effectively. In addition, some recent studies have made an effort to incorporate clinical variables, like different OCD symptom dimensions or treatment response profiles, which may help improve model relevance and accuracy. At the same time, it is worth being cautious. Not all increases in accuracy reflect genuine methodological progress, some may result from small sample sizes, lack of external validation, or overly optimistic performance metrics. These issues should be taken into account when interpreting trends over time.

Accuracy reporting

Reviewed studies employed various methods to report classification accuracy (cf., Table 7), making it difficult to identify the best-performing model. Differences in validation approaches significantly influenced reported accuracies. For example, a 1D CNN-LSTM model achieved 90.88% accuracy during cross-validation but dropped to 75% when tested on an external dataset. This highlights the need for standardized accuracy metrics and validation protocols. Since most studies reported overall accuracy, we used this measure for comparison (cf., Fig. 3). Due to the generally small sample sizes, analyzing additional performance metrics was not feasible. In general, deep learning models tended to outperform classical machine learning approaches.

Fig. 3
figure 3

Geographical Distribution of Studies

Machine learning models

According to Fig. 4, most articles used classical machine learning models. The most common model was SVM (used in 5 articles). Among deep learning models, ANN appeared in 3 articles. Tree-based models like RF, C5.0, and XGBoost were also common (5 articles) (Fig. 5).

Fig. 4
figure 4

Distribution of Reported Classification Accuracies for OCD Groups - Although accuracy is a common measure in studies, the difference in comparison groups (for example, healthy control group versus clinical control group) creates big differences. Thus, we must be coutious when comparing performance between studies

Fig. 5
figure 5

Hierarchy of Machine Learning Methods in OCD-EEG Studies

Interpretability methods

None of the reviewed studies employed modern interpretability methods such as SHAP or LIME. The authors recommend that future research incorporate these techniques to enhance the transparency of machine learning models, facilitating their translation into clinical tools like neurofeedback or electrode placement guidance for brain stimulation. While three studies attempted to interpret EEG features using statistical tests which is a positive step, these approaches are insufficient on their own to fully elucidate model decision-making.

Methodological comparison across studies using a shared dataset

Four studies reviewed in this paper (Park et al., 2021; Ahmed et al., 2024; Basak et al., 2023; and Mukherjee et al., 2024) all used the same dataset from Park 2021. This shared data gives a special chance to compare the methods used in each study. Because the data is the same, we can better see how model design, data processing, and validation affect the results. The accuracy reported in these studies is very different, from 74.52 to 100%. This shows that the choice of model and how complex it is can change the results a lot. For example, Mukherjee et al. used a deep model with Autoencoder and PLSTM and reported 100% accuracy. But they only used internal validation (cross-validation), and did not test the model on new data. So, this high result may not be real in practice and could mean overfitting. On the other hand, the simpler model Elastic Net used by Park et al. showed 74.52% accuracy, but it may generalize better because it is simple and clear.

To study how data preprocessing affects model results, we tried to find the same model in two papers: Ahmed and Basak. Both used LSTM models, but with different preprocessing. Ahmed used filtering, standardization, and coding. Basak used downsampling and moving average. But they used different ways to report results (Ahmed used direct accuracy, Basak used F1-Score), and they used different validation methods (cross-validation vs. train/test split). This made it hard to compare their results fairly. Also, Mukherjee et al. did not give information about preprocessing, so we could not include it in our data table. Because of these problems, we cannot clearly say how preprocessing affects model results in these studies.

All four studies used internal validation. None of them tested their models on new or outside data. Also, no study said clearly if their data split was based on subjects or just on EEG segments (epochs). This is important, because if data from the same person is in both training and testing, the result may look better than it really is. Without outside testing and clear reporting, it is hard to compare the models. We suggest that future studies use external validation and split data by subject, so we can understand how well models work in the real world. Using the same dataset across studies can also help us compare methods better, and this should be done more in future reviews.

Discussion

In this systematic review, we aimed to review studies that have been attempted to classify OCD compared to healthy or other psychiatric and neurological conditions using EEG and with an application of machine learning and deep learning techniques. In our review, we witnessed that studies usually do not focus on one or two disorders but focus on large categories of disorders. It is suggested that e.g., journals, as topic planners for special issues, should encourage topics with more clinical use while respecting basic research, like studies that are more interpretable, not just research with higher accuracy or more disorders. These types of reports, only with limited information provided lead to misundersatidng and misinterpretations given the complexity if the disorder and associated morbidities. In recent years, with an advancement of AI, more researchers are interested to distinguish neuropsychiatric conditions against healthy condition and to further develop precision medicine and individualise treatments. Although the small number of studies and the high heterogeneity of extracted data limit quantitative analysis, this variability highlights a critical issue: the lack of internationally standardized research designs and methodological perspectives to efficiently investigate the differences in detection of mental disorders and to predict the onset or monitirng the treatment such as phramocotherapy. The absence of such standards hampers the reproducibility of findings. In this section, we first discuss the current state of the field based on the reviewed studies, followed by future directions where we propose a practical framework. This framework aims to address existing challenges, improve study repeatability, and ultimately enable meaningful meta-analyses.

Current status

Accuracy paradox and high heterogeneity among the included studies

While high accuracy in a computational model would first be thought of as an achievement, the problem is that extremely high accuracy levels can occur from overfitting or from an oversmall sample. In the study by Mukherjee et al. (2024), the PsyNet model with Autoencoder + PLSTM architecture was 100% accurate, but since the sample size was very low (only 46 participants) and there was no external validation, the genralisability of results must be dealt with cautious. In studies using external validation — for example, Farhad et al., 2024 — the accuracy decreased from 90.88% in cross-validation to 75% in external validation, clearly showing that how accuracy was reported had a great impact on the percentage reported. Future work, therefore, should focus on comparing various models and evaluate the extent to which higher accuracies are observed, while also achieving reproducibility, uniform reporting of accuracy, representative sample size, and transparency of method should be considered in order to provide the solid groundwork for progressing EEG-ML as useful clinical tools. More specifically, considering that OCD is a heterogenous neuropsychiatric disorder with other morbidities or variety of symptoms severity, require more indepth investigations by including demographic information of patients and aged matched healthy individuals, medication stutus, years of illness, types of treatments and behavioral and cognitive components associated with both healthy and OCD patients. These considerations would better contribute to our understanding about the disorder and neurobehavioral and cognitive dimensions impaired due to the symptoms and the history of individuals. This further enables clinicians to develop an evidenced-based therapeutic models for the OCD patients depending on their age and with higher precision. A key contributor to the high heterogeneity observed across EEG-based machine learning studies in OCD is the significant variation in clinical, methodological, and technical approaches. Clinically, participant characteristics differ widely between studies. These include variations in symptom severity, diagnostic procedures, comorbidity profiles, and treatment history. While some studies used structured diagnostic interviews such as the MINI or Y-BOCS, others relied on broader psychiatric evaluations or failed to specify their diagnostic criteria. Treatment status also varied: some studies included drug-naïve participants, whereas others examined individuals undergoing pharmacological interventions such as SSRIs or antipsychotics—factors known to influence neural activity and EEG signatures.

From a technical perspective, there is a lack of standardization in EEG acquisition and processing protocols. Studies differed in their choice of electrode montages, frequency bands of interest, filtering techniques, and resting-state conditions (e.g., eyes open vs. eyes closed). These discrepancies inevitably affect the recorded signal characteristics and complicate cross-study comparisons. Additionally, preprocessing pipelines and feature extraction strategies were inconsistent, ranging from handcrafted spectral features to deep learning-based representations. Machine learning models also varied substantially—not only in algorithmic choice (e.g., SVM, ANN, LSTM, CNN-GRU, XGBoost), but also in how model performance was evaluated (e.g., accuracy, F1 score, AUC, balanced accuracy), and whether the evaluation used cross-validation, external validation, or subgroup analysis.

Performance outcomes reflect this diversity. For instance, recent studies such as Mukherjee et al. (2024) and Ahmed et al. (2024) reported remarkably high classification accuracies (up to 100% and 96.83%, respectively), while others like Gour et al. (2023) and Erguzel et al. (2015) reported more modest results (e.g., 63.21% and 56.96%). These discrepancies cannot be solely attributed to model quality but also reflect underlying variability in dataset composition, preprocessing rigor, and evaluation standards. Studies with external validation, such as Farhad et al. (2024), often report lower accuracies (e.g., 75.00%), underscoring the challenges of generalizability in real-world settings.

Overall, the primary source of heterogeneity seems to lie in the lack of standardization in both clinical characterization and EEG processing workflows. Without consistent approaches to defining OCD subtypes, accounting for comorbidities, or applying harmonized preprocessing and modeling techniques, it remains difficult to make meaningful comparisons or draw robust conclusions across studies. Future research should prioritize the adoption of standardized protocols, greater transparency in reporting methodological details, and inclusion of metadata related to clinical profiles and signal processing steps. These measures are critical for improving reproducibility and advancing the clinical utility of EEG-based models in OCD classification.”

Continued preference for traditional models and trends in deep learning and hybrid architectures

Despite the increasing attention toward modern machine learning techniques, traditional models still dominate EEG-based classification studies. This ongoing reliance can be attributed to several practical reasons: classical models such as SVM, and tree-based algorithms are computationally efficient, require less training data, and are more familiar to researchers. SVMs, in particular, have remained widely used due to their ability to handle high-dimensional feature spaces and employ nonlinear kernels for detecting complex patterns. Similarly, tree-based models have remained popular because they offer built-in dimensionality reduction through ensemble strategies like bagging, while also providing interpretable outputs through feature importance rankings.

While deep learning models were less frequently used in earlier studies, there has been a noticeable increase in their adoption in more recent literature (2023–2025). This shift reflects a gradual movement toward more sophisticated modeling techniques that can better capture the dynamic and high-dimensional nature of EEG signals. Among these, LSTM networks and their variants such as Bi-directional LSTM (Bi-LSTM) have emerged as the most commonly applied architectures, representing 30.8% of all deep learning approaches observed. These models are particularly well-suited for EEG data, as they are designed to model sequential dependencies and preserve important temporal information over time.

The suitability of LSTM-based models is especially relevant in the context of OCD research, where datasets tend to be relatively small. Unlike many deep learning architectures that demand large-scale data, LSTM models have shown greater resilience in smaller datasets, making them a practical and powerful tool for identifying temporal dynamics and subtle neurophysiological patterns often missed by traditional approaches. Their increasing usage signals a growing confidence in deep learning’s potential to enhance precision and interpretability in EEG-based OCD classification. Importantly, LSTMs are also more robust when dealing with small datasets—a recurring challenge in OCD-related EEG research. In machine learning, a common rule of thumb suggests that at least 10 samples per feature are needed to avoid overfitting. Given that EEG studies typically extract hundreds of features—e.g., from 19 electrodes across 5 frequency bands—a dataset with fewer than 100 participants is at high risk of overfitting. Yet, only seven OCD studies in this domain explicitly report sample sizes, and none include more than 95 OCD patients. This further justifies the use of LSTM-based models, which tend to perform better under such constraints compared to other deep learning architectures. As datasets grow and become more standardized, this trend toward deep learning is likely to accelerate.

Additionally, choosing between using raw signals or extracted features for EEG-ML in OCD is a complex problem. To our knowledge, it is not possible to say clearly if raw signals or extracted features are better. Raw signals, especially when given to LSTM or Transformer, may help models understand time series, sequences, and nonlinear features but may also include more noise. Using extracted features can reduce noise but hide time sequence data. We think both views are valuable, but it is recommended that if using extracted features, high beta and gamma bands should be included, because research has shown these bands are valuable for OCD diagnosis [10, 37,38,39]. Studies not analyzing these bands may miss key information, especially changes in e.g., associated with clinicl symptoms, monitoring medication effect and associated improvement.

Future directions: building toward reproducibility and clinical translation

To move EEG-based OCD research closer to clinical applications, future efforts must focus on consistency, transparency, and broader population relevance. One major challenge in this field is the lack of standardization, which makes it difficult to compare findings across studies or replicate results. Without consistency in methods and clarity in interpretation, promising discoveries may struggle to translate into clinical practice.

A notable observation from current literature is the geographic concentration of studies, with a majority carried out in Asian populations. While valuable, this regional focus may limit how well the findings apply to more diverse global populations. Expanding the use of publicly accessible EEG datasets that include a broader cultural and demographic range could help overcome this limitation.

Establishing a large-scale, open-access EEG repository with standardized protocols and data from varied populations would be a key step forward. Such a resource would not only strengthen reproducibility but also support more robust cross-study comparisons. This approach mirrors the success of the Human Connectome Project in the domain of fMRI, and serves as a model worth adapting for EEG research. Additionally, open datasets can help reduce the financial burden on individual research teams, foster collaboration, and accelerate the pace of innovation. They also promote transparency and allow the community to test and validate findings on independent data an essential requirement for clinical acceptance.

In sum, creating shared frameworks and accessible resources will be crucial for ensuring that findings in this field are not only scientifically sound but also relevant and usable in real-world clinical settings. Despite this need, an ethical framework need to be proposed by the scientific community due to the neurobiological datasets shared from human patients. A standardized broader framework that if not globally, but continent-wise is important. This would enable researchers to collaboratively share and investigate various dimensions of data based on their expertise and advance the field in a more efficient and constructive manner while respecting ethical concerns and regional and international data privacy law.

Data preprocessing

The reviewed literature demonstrates considerable variability in EEG preprocessing methods. We recommend the use of the COBIDAS-MEEG toolbox, which includes the following standardized steps: (1) detection and removal of bad channels, (2) artifact cleaning using techniques such as Independent Component Analysis (ICA) to remove eye and cardiac artifacts, (3) high-pass and low-pass filtering, (4) signal segmentation, (5) baseline correction, (6) transformation of electrode references, and (7) spectral transformation for frequency analysis. We also recommend furture research to test the differences across the same datasets when bad channels are interpolated and not, as in our opinion, noises might be part of the clinical symtpms. To ensure reproducibility, each step should be performed using consistent parameters and reported in exhaustive detail. Strict adherence to this standardized protocol will enhance methodological comparability and repeatability across studies.

Reporting accuracy

The reviewed studies exhibit considerable variability in how accuracy is reported. To address this issue, we recommend adhering to the TRIPOD guidelines (Transparent Reporting of a Multivariable Prediction Model) and encouraging the publication of full model training code, either within the paper or on platforms such as GitHub. Combining TRIPOD with the COBIDAS-MEEG standards for EEG preprocessing can establish a robust scientific foundation for the standardization of EEG-based machine learning research. This harmonization will facilitate more meaningful comparisons across studies and accelerate clinical adoption of predictive models. We further propose that international consortia should develop a specialized TRIPOD-EEG extension to better suit the unique aspects of EEG-based machine learning studies.

Model interpretation

Many models have been treated as “black boxes” without sufficient interpretation. Identifying which biomarkers, electrodes, or EEG waveforms are critical for OCD diagnosis and it is essential not only for accurate diagnosis but also for guiding treatment and precision medicine. Currently, interpretability is lacking, which undermines clinical trust in these models and limits the development of applications such as tES electrode placement and neurobiofeedback using feedback loops. As a result, most studies focused solely on binary diagnosis, restricting their potential impact on treatment strategies. Successful examples, such as FDA-approved neurofeedback devices for anxiety, demonstrate that EEG biomarkers can underpin non-invasive diagnostic and therapeutic tools.

While we encourage the integration of SHAP and LIME to enhance model transparency, it is important to recognize the practical considerations when applying these methods to EEG data. First, the effectiveness of these techniques is heavily dependent on the quality and relevance of input features. EEG data, often high-dimensional and noisy, can pose challenges in feature extraction, potentially reducing the interpretability of model outputs. Second, SHAP and LIME assume some degree of model agnosticism or structural compatibility—certain deep learning architectures, especially those working directly on raw EEG time-series data, may require preprocessing steps such as dimensionality reduction or segmentation to produce meaningful explanations. Third, interpreting the resulting importance scores in a neurophysiological context is non-trivial, as EEG features may not map directly to intuitive biomarkers without further domain knowledge or validation. Thus, while SHAP and LIME hold promise for improving transparency and guiding clinical applications (e.g., neurofeedback electrode placement or TES targeting), their implementation must be carefully tailored to the data type and clinical goals. Future studies should aim to validate the interpretability outputs with domain experts and explore hybrid strategies combining model-specific interpretability with domain-relevant visualization tools.

Such interpretability tools could serve as a bridge between machine learning model outputs and actionable clinical insights, such as identifying relevant biomarkers for individualized tES targeting or optimizing neurofeedback protocols. Their integration could significantly advance trust and usability in clinical environments, provided their methodological constraints are transparently addressed.

Overall, the lack of a clear conceptual and methodological connection between the reviewed findings and the specific symptoms of OCD makes it challenging to interpret the results in a way that is clinically useful. This gap seems to reflect the nature of the research questions often asked in these studies, which are frequently shaped by researchers from technical or engineering backgrounds. As a result, many of the studies tend to focus heavily on algorithmic or signal processing aspects, without anchoring their analyses in established clinical knowledge about OCD or in its underlying neurobiology. For example, few studies link their EEG findings to known patterns in OCD, such as altered theta or alpha activity, or consider how these patterns might differ across symptom types like contamination fears or compulsive checking. Similarly, the possibility of using EEG markers to predict responses to treatments such as SSRIs or cognitive behavioural therapy is not addressed. Without incorporating these clinical aspects, the research risks making broad, generalized claims that may not translate meaningfully to OCD. This lack of integration reduces the potential of EEG-ML methods to inform clinical practice or deepen understanding of OCD specifically. Going forward, studies, clinicians and patients involved would benefit from incorporating clinical variables such as symptom subtype, severity, or treatment history into their models and analyses. Aligning methodological development with clinical and theoretical knowledge about OCD could help the field move toward more targeted and informative findings, rather than remaining at the level of general critiques or technical demonstrations.

From challenge to vision

The greatest challenge identified in this review is the significant heterogeneity across the current literature, which complicates the identification of common features. To address this, we propose a three-stage framework (cf., Fig. 6) designed to guide researchers toward greater standardization and consistency by adhering to established protocols and addressing existing research gaps.

Fig. 6
figure 6

Three-stage standardization framework to reduce methodological heterogeneity and reduce research gap in EEG-based machine learning studies for OCD

Towards clinical translation

To move EEG-based machine learning from research to clinical settings, three translation gaps must be filled:

  1. 1.

    Regulatory readiness: Models must meet repeatability criteria in Fig. 6 to be homogeneous and FDA-reviewable.

  2. 2.

    Interpretability integration: Combine SHAP/LIME with models for better electrode placement in neurofeedback and TES, and better analysis of gamma and high beta waves.

  3. 3.

    Comorbidity management: Develop hybrid EEG-fMRI biomarkers to differentiate OCD from e.g., Depression/ADHD. Pilot studies using portable EEG headsets can test real-world feasibility.

Limitations

Despite conducting a thorough search across four major databases (PubMed, Scopus, IEEE, and Web of Science), and adhering to PRISMA guidelines, only 11 studies met our inclusion criteria. Considerable heterogeneity was observed across multiple aspects, including EEG recording protocols, preprocessing techniques, feature extraction methods, machine learning algorithms, and accuracy reporting. These challenges led to two primary conclusions: first, there is a critical need for standardized protocols and frameworks for EEG data acquisition and machine learning reporting; second, significant research gaps exist regarding model interpretability and the differentiation of OCD subtypes. Recognizing these limitations provides valuable insights that can guide future research toward greater reproducibility, enable more robust meta-analyses, and accelerate the translation of EEG-based machine learning approaches from experimental studies to clinical applications in OCD.

Future computationl and clinical direction

While the reviewed studies suggested promising avenues for using EEG and machine learning models in classifying OCD, the evidence base remains preliminary. Most included studies suffer from small sample sizes, lack of consistent external validation, and limited or absent interpretability methods. Therefore, translating these tools into clinical practice including regulatory milestones such as FDA approval or application in neurofeedback and tES targeting remains a long-term goal rather than an immediate prospect. Clinical utility is contingent on the development of standardized protocols, consistent reporting practices, and rigorous validation frameworks. Until these conditions are met, findings should be interpreted cautiously, especially when reported in terms of diagnostic or therapeutic utility such as tES and neuro-biofeedback interventions.

We suggest future studies to further consider following points during investigations:

As outlined in clinical diagnostic guidelines, OCD frequently co-occurs with a range of other neurodevelopmental and psychiatric conditions. When building computational models for OCD diagnosis, it’s important to account for this overlap. Brain activity patterns in individuals with OCD can be influenced by comorbidities such as anxiety disorders, ADHD, PTSD, tic disorders, and mood disturbances. These variations, shaped by the diverse clinical profiles seen in OCD, must be carefully considered to ensure that diagnostic models are accurate and generalizable.

Variations in OCD symptoms, treatment type, timing, and response to the clinical efforts also present key considerations for modeling. Pharmaceutical interventions, the stage of illness (e.g., early onset, chronic OCD), timing of symptom presentation, and the trajectory of therapeutic response all influence brain dynamics in meaningful ways. In particular, individuals with treatment-resistant OCD—an underrepresented group in many studies—may show distinct neurophysiological patterns. Including data from such cases can enhance predictive models and help tailor interventions such as cognitive-behavioral therapy, pharmacotherapy, neurofeedback, or neuromodulation techniques like transcranial magnetic stimulation.

Demographic variables, such as age and gender, also play a significant role in shaping OCD’s neural profile. For example, pediatric-onset OCD may present differently from adult-onset cases, and symptom expression often varies between males and females. Diagnostic models should be refined by integrating such demographic distinctions along with clinical factors like symptom severity, compulsivity levels, insight, and the presence of comorbid behaviors such as avoidance or sleep disruption. These parameters help create a more nuanced and individualized diagnostic approach.

To better differentiate OCD-related brain activity from healthy patterns, statistical analysis of extracted features should be complemented by more advanced signal processing techniques. Methods such as power spectral density (PSD) analysis and functional connectivity mapping can uncover subtle but meaningful distinctions in resting-state activity. Focusing on specific frequency bands and relevant brain regions those found to show altered communication in OCD can improve classification accuracy and strengthen the diagnostic relevance of these computational tools.

In addition, some studies in this review reported very high accuracies or made claims regarding clinical applicability without adequate supporting evidence. For instance, the use of terms like “diagnostic tool” or “clinical prediction” was sometimes made in the absence of external validation, interpretability analysis, or outcome-based evaluation. To improve scientific transparency and prevent overinterpretation:

  • Future authors should clearly state whether their findings are exploratory or validated.

  • Claims about clinical utility should be qualified unless supported by prospective trials, larger sample sizes, or real-world evaluations.

  • Interpretability tools (e.g., SHAP, Grad-CAM) should accompany performance metrics to support clinical translation.

  • Studies should avoid equating high cross-validation accuracy with diagnostic readiness.

Finally, dimensionality reduction and advanced modeling techniques offer promising pathways to uncover deeper patterns within the data. Techniques like t-distributed stochastic neighbor embedding (t-SNE) can be used to explore whether neural data clusters by OCD symptom severity or subtype. Similarly, using architectures like autoencoders, transformers, or unsupervised learning methods can help reveal hidden structure in EEG data, enhance model interpretability, and support the development of more robust, data-driven diagnostic frameworks for OCD.

Conclusion

This systematic review highlights the emerging yet highly heterogeneous field of EEG-based machine learning for classifying OCD. Traditional machine learning models such as SVM and Random Forest have been commonly used, while more recent studies increasingly apply deep learning architectures like LSTM and CNN-LSTM. Reported classification accuracies varied widely, with studies employing external validation demonstrating more reliable results. These findings underscore the urgent need for standardized protocols in EEG data acquisition, preprocessing, model reporting, and interpretability. Our proposed three-step standardization framework offers a practical roadmap to enhance reproducibility and lay the foundation for future meta-analyses. While EEG-based machine learning holds potential for aiding OCD diagnosis, the current state of the field does not yet support clinical deployment. Future research should prioritize large, well-characterized samples, standardized protocols, interpretability tools, and outcome validation to move toward genuine clinical readiness.

Data availability

No datasets were generated or analysed during the current study.

Abbreviations

SCZ:

Schizophrenia

MDD:

Depression disorder

BD:

Bipolar disorder

PD:

Panic disorder

SAD:

Social anxiety disorder

OCD:

Obsessive-compulsive disorder

AUD:

Alcohol use disorder

BA:

Behavioral addiction

PTSD:

Post-traumatic stress disorder

ASD:

Acute stress disorder

AD:

Adaptation disorder

SUD:

Substance use disorder

ADHD:

Attention-deficit/hyperactivity disorder

OUD:

Opiate use adective

HC:

Healthy control

TTM:

Trichotillomania

SMC:

Subjective memory complaints

References

  1. Stein DJ, et al. Obsessive–compulsive disorder. Nat Reviews Disease Primers. 2019;5:52. https://doi.org/10.1038/s41572-019-0102-3.

    Article  PubMed  Google Scholar 

  2. Angst J, et al. Obsessive-compulsive severity spectrum in the community: prevalence, comorbidity, and course. Eur Arch Psychiatry Clin NeuroSci. 2004;254:156–64. https://doi.org/10.1007/s00406-004-0459-4.

    Article  PubMed  Google Scholar 

  3. Hollander E, et al. Psychosocial function and economic costs of Obsessive-Compulsive disorder. CNS Spectr. 1998;3:48–58. https://doi.org/10.1017/S1092852900007239.

    Article  Google Scholar 

  4. Selwood L. Obsessive compulsive disorder. SA Pharm J. 2014;81:30–2.

    Google Scholar 

  5. Hezel DM, Rose SV, Simpson HB. Delay to diagnosis in OCD. J Obsessive-Compulsive Relat Disorders. 2022;32:100709. https://doi.org/10.1016/j.jocrd.2022.100709.

    Article  Google Scholar 

  6. Porandla K, Nagpal R, Garg S. Expert opinion on challenges in the diagnosis and treatment of OCD in India. Int J Res Med Sci. 2024;13:522–31. https://doi.org/10.18203/2320-6012.ijrms20244166.

    Article  Google Scholar 

  7. Glazier K, Swing M, McGinn LK. Half of obsessive-compulsive disorder cases misdiagnosed: vignette-based survey of primary care physicians. J Clin Psychiatry. 2015;76:7995.

    Article  Google Scholar 

  8. Chasson GS, Williams MT, Davis DM, Combs JY. Missed diagnoses in African Americans with obsessive-compulsive disorder: the structured clinical interview for DSM-IV axis I disorders (SCID-I). BMC Psychiatry. 2017;17:258. https://doi.org/10.1186/s12888-017-1422-z.

    Article  PubMed Central  PubMed  Google Scholar 

  9. Hirschtritt ME, Bloch MH, Mathews CA. Obsessive-Compulsive disorder: advances in diagnosis and treatment. JAMA. 2017;317:1358–67. https://doi.org/10.1001/jama.2017.2200.

    Article  PubMed  Google Scholar 

  10. Karadag F, et al. QUANTITATIVE EEG ANALYSIS IN OBSESSIVE COMPULSIVE DISORDER. Int J Neurosci. 2003;113:833–47. https://doi.org/10.1080/00207450390200963.

    Article  PubMed  Google Scholar 

  11. Tot Ş, Özge A, Çömelekoğlu Ü, Yazici K, Bal N. Association of QEEG findings with clinical characteristics of OCD: evidence of left frontotemporal dysfunction. Can J Psychiatry. 2002;47:538–45. https://doi.org/10.1177/070674370204700605.

    Article  PubMed  Google Scholar 

  12. Bell MAand, Cuevas K, Using. EEG to study cognitive development: issues and practices. J Cognition Dev. 2012;13:281–94. https://doi.org/10.1080/15248372.2012.691143.

    Article  CAS  Google Scholar 

  13. Bearden STand, Nay LB. Utility of EEG in differential diagnosis of adults with unexplained acute alteration of mental status. Am J Electroneurodiagn Technol. 2011;51:92–104. https://doi.org/10.1080/1086508X.2011.11079807.

    Article  Google Scholar 

  14. Anwar T, Rehmat N, Naveed H. in 2021 43rd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC). 2025–2029.

  15. Rivera MJ, Teruel MA, Maté A, Trujillo J. Diagnosis and prognosis of mental disorders by means of EEG and deep learning: a systematic mapping study. Artif Intell Rev. 2022;55:1209–51. https://doi.org/10.1007/s10462-021-09986-y.

    Article  Google Scholar 

  16. Sajno E, et al. Machine learning in biosignals processing for mental health: A narrative review. Front Psychol. 2023;13–2022. https://doi.org/10.3389/fpsyg.2022.1066317.

  17. Keizer AW. Standardization and personalized medicine using quantitative EEG in clinical settings. Clin EEG Neurosci. 2021;52:82–9. https://doi.org/10.1177/1550059419874945.

    Article  PubMed  Google Scholar 

  18. Alreshidi IM, Moulitsas I, Jenkins KW. in Proceedings of the 6th International Conference on Advances in Artificial Intelligence 29–39Association for Computing Machinery, Birmingham, United Kingdom, (2023).

  19. Ciodaro G, Najafabadi AJ, Godde B. in 2020 IEEE Signal Processing in Medicine and Biology Symposium (SPMB). 1–6.

  20. Bhaskarachary C, Najafabadi AJ, Godde B. in 2020 IEEE Signal Processing in Medicine and Biology Symposium (SPMB). 1–4.

  21. Bhaskarachary C, Jahanian Najafabadi A, Godde B. Machine Learning Supervised Classification Methodology for Autism Spectrum Disorder based on Resting-State Electroencephalography (EEG) Signals. (2021).

  22. Najafabadi AJ, Bagh K. in 2023 IEEE Signal Processing in Medicine and Biology Symposium (SPMB). 1–6.

  23. Najafabadi AJ, Bagh K. in 2024 IEEE Signal Processing in Medicine and Biology Symposium (SPMB). 1–7 (IEEE).

  24. Najafabadi AJ, Bagh K. Deep learning analysis approach for major depressive disorder in children and adolescents. Preprint (2025).

  25. Craik A, He Y, Contreras-Vidal JL. Deep learning for electroencephalogram (EEG) classification tasks: a review. J Neural Eng. 2019;16:031001.

    Article  PubMed  Google Scholar 

  26. Vimbi V, Shaffi N, Mahmud M. Interpreting artificial intelligence models: a systematic review on the application of LIME and SHAP in alzheimer’s disease detection. Brain Inf. 2024;11:10. https://doi.org/10.1186/s40708-024-00222-1.

    Article  Google Scholar 

  27. Mane D, et al. Unlocking machine learning model decisions: A comparative analysis of LIME and SHAP for enhanced interpretability. J Electr Syst. 2024;20:1252–67.

    Article  Google Scholar 

  28. Hasan MM. Understanding model predictions: a comparative analysis of SHAP and LIME on various ML algorithms. J Sci Technological Res. 2023;5:17–26.

    Google Scholar 

  29. Zhao Z et al. Predicting treatment response of repetitive transcranial magnetic stimulation in major depressive disorder using an explainable machine learning model based on electroencephalography and clinical features. Biol Psychiatry Cogn Neurosci Neuroimaging, S2451-9022 (2425) 00059–X.

  30. Islam MS, Hussain I, Rahman MM, Park SJ, Hossain MA. Explainable artificial intelligence model for stroke prediction using EEG signal. Sensors. 2022;22:9859.

    Article  PubMed Central  PubMed  Google Scholar 

  31. Sallum LF et al. Revealing patterns in major depressive disorder with machine learning and networks. medrxiv, 2024.2006. 2007.24308619 (2024).

  32. McLoughlin G, Makeig S, Tsuang MT. In search of biomarkers in psychiatry: EEG-based measures of brain function. Am J Med Genet Part B: Neuropsychiatric Genet. 2014;165:111–21.

    Article  Google Scholar 

  33. Kopřivová J, et al. EEG source analysis in obsessive-compulsive disorder. Clin Neurophysiol. 2011;122:1735–43. https://doi.org/10.1016/j.clinph.2011.01.051.

    Article  PubMed  Google Scholar 

  34. Özçoban MA, Kara S, Tan O, Aydın S. in 2014 18th National Biomedical Engineering Meeting. 1–3.

  35. Yu J, et al. Symptom provocation in obsessive–compulsive disorder: A voxel-based meta-analysis and meta-analytic connectivity modeling. J Psychiatr Res. 2022;146:125–34. https://doi.org/10.1016/j.jpsychires.2021.12.029.

    Article  PubMed  Google Scholar 

  36. Rauch, S. L.et al. Functional Magnetic Resonance Imaging Study of Regional Brain Activation During Implicit Sequence Learning in Obsessive–Compulsive Disorder. Biological Psychiatry 61, 330–336 (2007). https://doi.org/10.1016/j.biopsych.2005.12.012.

  37. Treu S, et al. A ventromedial prefrontal dysrhythmia in obsessive-compulsive disorder is attenuated by nucleus accumbens deep brain stimulation. Brain Stimulation: Basic Translational Clin Res Neuromodulation. 2021;14:761–70. https://doi.org/10.1016/j.brs.2021.04.028.

    Article  Google Scholar 

  38. Newson JJ, Thiagarajan TC. EEG frequency bands in psychiatric disorders: a review of resting state studies. Front Hum Neurosci. 2019;12:521.

    Article  PubMed Central  PubMed  Google Scholar 

  39. Velikova S, et al. Dysfunctional brain circuitry in obsessive–compulsive disorder: source and coherence analysis of EEG rhythms. NeuroImage. 2010;49:977–83. https://doi.org/10.1016/j.neuroimage.2009.08.015. https://doi.org/https://doi.org/.

    Article  PubMed  Google Scholar 

Download references

Funding

Open Access funding enabled and organized by Projekt DEAL. Authors declare no financial support for this systematic study.

Author information

Authors and Affiliations

  1. Department of Cognitive Sciences, Faculty of Psychology and Educational Sciences, Shiraz University, Shiraz, Iran

    Mahdi Naderi

  2. Department of Cognitive Neuroscience, Faculty of Biology, Bielefeld University, Bielefeld, Germany

    Amir Jahanian-Najafabadi

Authors

  1. Mahdi Naderi
  2. Amir Jahanian-Najafabadi

Contributions

MN: Conceptualization and design of the study, data extraction, quality assessment, writing; AJN: Conceptualization and design of the study, data extraction, quality assessment, methodology, writing, reviewing, and editing. All authors reviewed the manuscript.

Corresponding author

Correspondence to Amir Jahanian-Najafabadi.

Ethics declarations

Ethics approval and consent to participate

Not applicable.

Competing interests

The authors declare no competing interests.

Clinical trial number

Not applicable.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Naderi, M., Jahanian-Najafabadi, A. A systematic review of EEG-based machine learning classifications for obsessive-compulsive disorder: current status and future directions. BMC Psychiatry 25, 854 (2025). https://doi.org/10.1186/s12888-025-07296-z

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12888-025-07296-z

Keywords

关于《A systematic review of EEG-based machine learning classifications for obsessive-compulsive disorder: current status and future directions》的评论


暂无评论

发表评论

摘要

### Summary of the Systematic Review on EEG-Based Machine Learning Classifications for Obsessive-Compulsive Disorder (OCD) **Authors:** Mahdi Naderi, Amir Jahanian-Najafabadi #### Abstract: The study reviewed existing literature to assess the application and potential future directions of Electroencephalography (EEG)-based machine learning methods in diagnosing and understanding Obsessive-Compulsive Disorder (OCD). The review aimed at evaluating current methodologies, their effectiveness, limitations, and possible advancements. **Keywords:** EEG; Machine Learning; OCD; Diagnosis; Neuroinformatics --- ### Introduction: Obsessive-Compulsive Disorder (OCD) is a neuropsychiatric condition characterized by recurrent obsessive thoughts and compulsive behaviors. Accurate diagnosis and understanding of OCD can be enhanced through the application of advanced technologies like Electroencephalography (EEG) combined with machine learning techniques. **Aims:** 1. To evaluate the current status of EEG-based machine learning methods in classifying OCD. 2. To identify limitations and challenges within existing studies. 3. To suggest future research directions for improving diagnostic accuracy and clinical utility. --- ### Methods: The review was conducted following PRISMA guidelines, with searches carried out on databases such as PubMed, IEEE Xplore, and Web of Science. Studies were included if they involved machine learning techniques applied to EEG data specifically for OCD classification or related purposes. Quality assessment was performed using standardized tools focusing on methodological rigor. --- ### Results: 12 articles met the inclusion criteria for review. Key findings include: #### Methodology & Techniques Used: - **EEG Data Collection:** Various methods were used, including resting-state and task-based EEG recordings. - **Machine Learning Algorithms:** Commonly applied algorithms included Support Vector Machines (SVM), Random Forests (RF), Neural Networks (NN), and Deep Learning models. #### Classification Performance: - Overall accuracy ranged from 60% to 95%, with deep learning methods generally outperforming traditional machine learning approaches. - Specificity and sensitivity varied widely, reflecting the complexity of OCD diagnosis. #### Challenges & Limitations: - **Data Quality:** Variability in EEG data quality across studies affected classification performance. - **Sample Size:** Limited sample sizes hindered generalizability of findings. - **Heterogeneity:** Significant heterogeneity among patient populations further complicates diagnostic consistency. --- ### Discussion: The review highlights the potential but also limitations of using machine learning on EEG data for OCD diagnosis. Key points include: 1. **Advancements in Machine Learning Techniques:** - Deep learning models have shown promising results, particularly in handling large and complex datasets. 2. **Standardization Needs:** - Standardizing EEG acquisition protocols and preprocessing methods is crucial to enhance comparability across studies. 3. **Integration of Clinical Data:** - Combining EEG data with clinical assessments can provide more robust diagnostic models. 4. **Clinical Utility:** - Future research should focus on translating these findings into practical clinical applications, such as early detection and personalized treatment planning. --- ### Conclusion: EEG-based machine learning holds significant promise for advancing OCD diagnosis and understanding. However, methodological challenges must be addressed to realize this potential fully. Standardization of methodologies, larger-scale studies with diverse populations, and integration with other clinical data are recommended. **Future Directions:** 1. Develop standardized protocols for EEG acquisition and processing. 2. Enhance data sharing initiatives to support collaborative research efforts. 3. Investigate the utility of machine learning models in predicting treatment response and patient outcomes. --- ### Funding: This study received no specific financial support. ### Ethics Approval & Consent to Participate: Not applicable. ### Competing Interests: The authors declare no competing interests. ### Reprints and Permissions: This article is licensed under a Creative Commons Attribution 4.0 International License, allowing for use, sharing, adaptation, distribution, and reproduction in any medium or format provided appropriate credit is given to the original author(s) and source, with links to the license included as specified. --- **Keywords:** EEG; Machine Learning; OCD; Diagnosis; Neuroinformatics