10.25673/122075">


Proceedings of International Conference on Applied Innovation in IT
2025/08/29, Volume 13, Issue 4, pp.117-123

A Deep Learning-Driven Filter Bank CSP Approach for Motor Imagery EEG Decoding


Reem Dheyaa Ismael, Rasha Abdalla Yousif, Wisal Saud Hussein, Rasha Ahmed Jelab, Mohamed Aktham Ahmed and Zaidoon Tareq Abdulwahhab


Abstract: Motor Imagery (MI)-based Brain-Computer Interfaces (BCIs) have garnered considerable interest for facilitating direct neurological control of external devices, especially in assistive and rehabilitative technologies. However, the proficient classification of non-stationary, low-amplitude EEG signals continue to remain a significant difficulty. This study introduces a hybrid framework that combines Filter Bank Common Spatial Pattern (FBCSP) with a 1D Convolutional Neural Network (CNN) to improve the classification of motor imagery signals. Electroencephalogram data were acquired from four participants utilizing a 14-channel Emotive EPOC headset during two-class motor imaging tasks (left versus right hand imagery). The EEG samples were bandpass filtered into three frequency sub-bands (8-12 Hz, 12-16 Hz, and 16-30 Hz), and to extract discriminative spatial features the CSP was applied to each band. These features were combined and normalized before being entered into a lightweight CNN model for classification. The model was trained with the Adam optimizer and evaluated using standard metrics. Subject-specific results showed high classification ability, with accuracy approaching 100% for some individuals and an average accuracy above 90% across most subjects. The proposed FBCSP + CNN pipeline effectively captures spatial-spectral patterns in EEG data while being computationally inexpensive, making it ideal for real-time BCI applications that use consumer-grade EEG sensors. These findings emphasize the utility of hybrid handcrafted-deep learning models in actual MI-BCI systems.

Keywords: Brain-Computer Interface, Motor Imagery, EEG, Filter Bank CSP, Convolutional Neural Network, Emotiv EPOC.

DOI: 10.25673/122075

Download: PDF

References:

  1. Y. Qiu, H. Liu, and M. Zhao, “A review of brain–computer interface-based language decoding: From signal interpretation to intelligent communication,” Applied Sciences, vol. 15, p. 392, 2025.
  2. Y. Wang, C. Jiang, and C. Li, “A review of brain–computer interface technologies: Signal acquisition methods and interaction paradigms,” arXiv preprint arXiv:2503.16471, 2025.
  3. S. Pérez-Velasco, D. Marcos-Martínez, E. Santamaría-Vázquez, V. Martínez-Cagigal, and R. Hornero, “Bridging motor execution and motor imagery BCI paradigms: An inter-task transfer learning approach,” Biomed. Signal Process. Control, vol. 107, p. 107834, 2025.
  4. I. Hameed, D. M. Khan, S. M. Ahmed, S. S. Aftab, and H. Fazal, “Enhancing motor imagery EEG signal decoding through machine learning: A systematic review of recent progress,” Comput. Biol. Med., vol. 185, p. 109534, 2025.
  5. R. A. Aljanabi, Z. Al-Qaysi, M. Ahmed, and M. M. Salih, “Hybrid model for motor imagery biometric identification,” Iraqi J. Comput. Sci. Math., vol. 5, pp. 1–12, 2024.
  6. M. Heim, F. Heinrichs, M. Hueppe, F. Nunez, A. Szameitat, M. Reuter, et al., “Real-time EEG-based BCI for self-paced motor imagery and motor execution using functional neural networks,” IEEE Access, 2025.
  7. Z. Al-Qaysi, A. Al-Saegh, A. F. Hussein, and M. Ahmed, “Wavelet-based hybrid learning framework for motor imagery classification,” Iraqi J. Electr. Electron. Eng., 2022.
  8. L. Zhao, Y. Liu, J. A. Gao, P. Ding, F. Wang, A. Gong, et al., “Visual imagery-based brain–computer interaction paradigms and neural encoding and decoding,” IEEE Trans. Hum.-Mach. Syst., 2025.
  9. Y. Xie and S. Oniga, “Classification of motor imagery EEG signals based on data augmentation and convolutional neural networks,” Sensors, vol. 23, p. 1932, 2023.
  10. A. Jiang, J. Shang, X. Liu, Y. Tang, H. K. Kwan, and Y. Zhu, “Efficient CSP algorithm with spatio-temporal filtering for motor imagery classification,” IEEE Trans. Neural Syst. Rehabil. Eng., vol. 28, pp. 1006–1016, 2020.
  11. Y. Lu and J. Chen, “Improved EEG-based emotion classification via Stockwell entropy and CSP integration,” Entropy, vol. 27, p. 457, 2025.
  12. M. I. Alam, M. M. Nuhash, A. Zihad, T. H. Nakib, and M. M. Ehsan, “Conventional and emerging CSP technologies and design modifications: Research status and recent advancements,” Int. J. Thermofluids, vol. 20, p. 100406, 2023.
  13. B. Khalid, A. Hassan, M. Yasin, M. Salman, M. F. U. Butt, W. Abdul, et al., “A triple-shallow CNN with genetic algorithm channel selection method for classifying EEG complex limb movements,” Biomed. Signal Process. Control, vol. 105, p. 107541, 2025.
  14. J. Lu, Y. Tian, Y. Zhang, Q. Z. Sheng, and X. Zheng, “LGL-BCI: A motor-imagery-based brain–computer interface with geometric learning,” Proc. ACM Interact. Mob. Wearable Ubiquitous Technol., vol. 9, pp. 1–28, 2025.
  15. Z. Al-Qaysi, M. Suzani, N. bin Abdul Rashid, R. D. Ismail, M. Ahmed, R. A. Aljanabi, et al., “Generalized time-domain prediction model for motor imagery-based wheelchair movement control,” Mesopotamian J. Big Data, pp. 68–81, 2024.
  16. Z. Al-Qaysi, M. Suzani, N. bin Abdul Rashid, R. D. Ismail, M. Ahmed, W. A. W. Sulaiman, et al., “A frequency-domain pattern recognition model for motor imagery-based brain–computer interface,” Appl. Data Sci. Anal., pp. 82–100, 2024.
  17. S. Chaudhary, S. Taran, V. Bajaj, and A. Sengur, “Convolutional neural network based approach towards motor imagery tasks EEG signals classification,” IEEE Sensors J., vol. 19, pp. 4494–4500, 2019.


    HOME

       - Conference
       - Journal
       - Paper Submission to Journal
       - Paper Submission to Conference
       - For Authors
       - For Reviewers
       - Important Dates
       - Conference Committee
       - Editorial Board
       - Reviewers
       - Last Proceedings


    PROCEEDINGS

       - Volume 13, Issue 4 (ICAIIT 2025)
       - Volume 13, Issue 3 (ICAIIT 2025)
       - Volume 13, Issue 2 (ICAIIT 2025)
       - Volume 13, Issue 1 (ICAIIT 2025)
       - Volume 12, Issue 2 (ICAIIT 2024)
       - Volume 12, Issue 1 (ICAIIT 2024)
       - Volume 11, Issue 2 (ICAIIT 2023)
       - Volume 11, Issue 1 (ICAIIT 2023)
       - Volume 10, Issue 1 (ICAIIT 2022)
       - Volume 9, Issue 1 (ICAIIT 2021)
       - Volume 8, Issue 1 (ICAIIT 2020)
       - Volume 7, Issue 1 (ICAIIT 2019)
       - Volume 7, Issue 2 (ICAIIT 2019)
       - Volume 6, Issue 1 (ICAIIT 2018)
       - Volume 5, Issue 1 (ICAIIT 2017)
       - Volume 4, Issue 1 (ICAIIT 2016)
       - Volume 3, Issue 1 (ICAIIT 2015)
       - Volume 2, Issue 1 (ICAIIT 2014)
       - Volume 1, Issue 1 (ICAIIT 2013)


    PAST CONFERENCES

       ICAIIT 2025
         - Photos
         - Reports

       ICAIIT 2024
         - Photos
         - Reports

       ICAIIT 2023
         - Photos
         - Reports

       ICAIIT 2021
         - Photos
         - Reports

       ICAIIT 2020
         - Photos
         - Reports

       ICAIIT 2019
         - Photos
         - Reports

       ICAIIT 2018
         - Photos
         - Reports

    ETHICS IN PUBLICATIONS

    ACCOMODATION

    CONTACT US

 

        

         Proceedings of the International Conference on Applied Innovations in IT by Anhalt University of Applied Sciences is licensed under CC BY-SA 4.0


                                                   This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License


           ISSN 2199-8876
           Publisher: Edition Hochschule Anhalt
           Location: Anhalt University of Applied Sciences
           Email: leiterin.hsb@hs-anhalt.de
           Phone: +49 (0) 3496 67 5611
           Address: Building 01 - Red Building, Top floor, Room 425, Bernburger Str. 55, D-06366 Köthen, Germany

        site traffic counter

Creative Commons License
Except where otherwise noted, all works and proceedings on this site is licensed under Creative Commons Attribution-ShareAlike 4.0 International License.