IJESRT. (I2OR), Publication Impact Factor: 3.785

Similar documents
Brain-Computer Interface (BCI)

A BCI Control System for TV Channels Selection

EEG Eye-Blinking Artefacts Power Spectrum Analysis

Feature Conditioning Based on DWT Sub-Bands Selection on Proposed Channels in BCI Speller

18th European Signal Processing Conference (EUSIPCO-2010) Aalborg, Denmark, August 23-27, GIPSA-lab CNRS UMR 5216

Common Spatial Pattern Ensemble Classifier and Its Application in Brain-Computer Interface

Reliable visual stimuli on LCD screens for SSVEP based BCI

Common Spatial Patterns 2 class BCI V Copyright 2012 g.tec medical engineering GmbH

Common Spatial Patterns 3 class BCI V Copyright 2012 g.tec medical engineering GmbH

Research Article. ISSN (Print) *Corresponding author Shireen Fathima

Decoding of Multichannel EEG Activity from the Visual Cortex in. Response to Pseudorandom Binary Sequences of Visual Stimuli

Smart Traffic Control System Using Image Processing

Keywords Separation of sound, percussive instruments, non-percussive instruments, flexible audio source separation toolbox

International Journal of Advance Research in Engineering, Science & Technology

BCI Autonomous Assistant System with Seven Tasks for Assisting Disable People

Evolutionary Brain Computer Interfaces

ECG Denoising Using Singular Value Decomposition

Reconstruction of Ca 2+ dynamics from low frame rate Ca 2+ imaging data CS229 final project. Submitted by: Limor Bursztyn

MindMouse. This project is written in C++ and uses the following Libraries: LibSvm, kissfft, BOOST File System, and Emotiv Research Edition SDK.

Environmental Controls Laboratory

Completing Cooperative Task by Utilizing EEGbased Brain Computer Interface

Research Article Music Composition from the Brain Signal: Representing the Mental State by Music

Singer Traits Identification using Deep Neural Network

Brain Computer Music Interfacing Demo

BRAIN-COMPUTER interface (BCI) systems have been

Skip Length and Inter-Starvation Distance as a Combined Metric to Assess the Quality of Transmitted Video

Distortion Analysis Of Tamil Language Characters Recognition

MUSI-6201 Computational Music Analysis

SOBI-RO for Automatic Removal of Electroocular Artifacts from EEG Data-Based Motor Imagery

Error Resilience for Compressed Sensing with Multiple-Channel Transmission

Work In Progress: Adapting Inexpensive Game Technology to Teach Principles of Neural Interface Technology and Device Control

Digital Correction for Multibit D/A Converters

Audio-Based Video Editing with Two-Channel Microphone

Reduction of Noise from Speech Signal using Haar and Biorthogonal Wavelet

HBI Database. Version 2 (User Manual)

International Journal of Advance Engineering and Research Development MUSICAL INSTRUMENT IDENTIFICATION AND STATUS FINDING WITH MFCC

OBJECTIVE EVALUATION OF A MELODY EXTRACTOR FOR NORTH INDIAN CLASSICAL VOCAL PERFORMANCES

Department of Electrical & Electronic Engineering Imperial College of Science, Technology and Medicine. Project: Real-Time Speech Enhancement

gresearch Focus Cognitive Sciences

BRAIN-ACTIVITY-DRIVEN REAL-TIME MUSIC EMOTIVE CONTROL

Music Tempo Classification Using Audio Spectrum Centroid, Audio Spectrum Flatness, and Audio Spectrum Spread based on MPEG-7 Audio Features

An Empirical Study on Identification of Strokes and their Significance in Script Identification

Instrument Recognition in Polyphonic Mixtures Using Spectral Envelopes

A Parametric Autoregressive Model for the Extraction of Electric Network Frequency Fluctuations in Audio Forensic Authentication

Space-Time-Frequency Bag of Words Models for Capturing EEG Variability: A Comprehensive Study

Artifact rejection and running ICA

INTER GENRE SIMILARITY MODELLING FOR AUTOMATIC MUSIC GENRE CLASSIFICATION

An Efficient Low Bit-Rate Video-Coding Algorithm Focusing on Moving Regions

Hardware/Software Components and Applications of BCIs

Removal Of EMG Artifacts From Multichannel EEG Signal Using Automatic Dynamic Segmentation

ERROR CONCEALMENT TECHNIQUES IN H.264 VIDEO TRANSMISSION OVER WIRELESS NETWORKS

Doubletalk Detection

the effects of monitor raster latency on VEPs and ERPs. and Brain-Computer Interface performance

Adaptive Key Frame Selection for Efficient Video Coding

A QUERY BY EXAMPLE MUSIC RETRIEVAL ALGORITHM

Real Time Bio-signal Acquisition System

Efficient Implementation of Neural Network Deinterlacing

IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING, VOL. 00, NO. 00,

Research Topic. Error Concealment Techniques in H.264/AVC for Wireless Video Transmission in Mobile Networks

UNIVERSITA DEGLI STUDI DI CATANIA Dipartimento di Ingegneria Elettrica Elettronica e dei Sistemi DIEES. Paola Belluomo

Bi-Modal Music Emotion Recognition: Novel Lyrical Features and Dataset

Topics in Computer Music Instrument Identification. Ioanna Karydi

Multiple-Window Spectrogram of Peaks due to Transients in the Electroencephalogram

Automatic Piano Music Transcription

Speech and Speaker Recognition for the Command of an Industrial Robot

Detection and demodulation of non-cooperative burst signal Feng Yue 1, Wu Guangzhi 1, Tao Min 1

Optimum Frame Synchronization for Preamble-less Packet Transmission of Turbo Codes

Investigation of Digital Signal Processing of High-speed DACs Signals for Settling Time Testing

Hidden Markov Model based dance recognition

An Introduction to the Spectral Dynamics Rotating Machinery Analysis (RMA) package For PUMA and COUGAR

A Parametric Autoregressive Model for the Extraction of Electric Network Frequency Fluctuations in Audio Forensic Authentication

Decision-Maker Preference Modeling in Interactive Multiobjective Optimization

ONE SENSOR MICROPHONE ARRAY APPLICATION IN SOURCE LOCALIZATION. Hsin-Chu, Taiwan

FPGA DESIGN OF CLUTTER GENERATOR FOR RADAR TESTING

Re: ENSC 370 Project Physiological Signal Data Logger Functional Specifications

Voice & Music Pattern Extraction: A Review

Region Adaptive Unsharp Masking based DCT Interpolation for Efficient Video Intra Frame Up-sampling

Extraction Methods of Watermarks from Linearly-Distorted Images to Maximize Signal-to-Noise Ratio. Brandon Migdal. Advisors: Carl Salvaggio

Study of White Gaussian Noise with Varying Signal to Noise Ratio in Speech Signal using Wavelet

Music Source Separation

Automatic Music Clustering using Audio Attributes

Figure.1 Clock signal II. SYSTEM ANALYSIS

Digital holographic security system based on multiple biometrics

... A Pseudo-Statistical Approach to Commercial Boundary Detection. Prasanna V Rangarajan Dept of Electrical Engineering Columbia University

Chord Classification of an Audio Signal using Artificial Neural Network

VivoSense. User Manual Galvanic Skin Response (GSR) Analysis Module. VivoSense, Inc. Newport Beach, CA, USA Tel. (858) , Fax.

Research Article Design and Analysis of a High Secure Video Encryption Algorithm with Integrated Compression and Denoising Block

THE importance of music content analysis for musical

Robert Alexandru Dobre, Cristian Negrescu

TERRESTRIAL broadcasting of digital television (DTV)

A Categorical Approach for Recognizing Emotional Effects of Music

Operating Bio-Implantable Devices in Ultra-Low Power Error Correction Circuits: using optimized ACS Viterbi decoder

Music BCI ( )

Multichannel Satellite Image Resolution Enhancement Using Dual-Tree Complex Wavelet Transform and NLM Filtering

Design of effective algorithm for Removal of Ocular Artifact from Multichannel EEG Signal Using ICA and Wavelet Method

Implementation of BIST Test Generation Scheme based on Single and Programmable Twisted Ring Counters

Design of a Mobile Brain Computer Interface-Based Smart Multimedia Controller

MUSICAL NOTE AND INSTRUMENT CLASSIFICATION WITH LIKELIHOOD-FREQUENCY-TIME ANALYSIS AND SUPPORT VECTOR MACHINES

Reducing tilt errors in moiré linear encoders using phase-modulated grating

Musical Instrument Identification Using Principal Component Analysis and Multi-Layered Perceptrons

Transcription:

[Kaushik, 4(8): Augusts, 215] ISSN: 2277-9655 IJESRT INTERNATIONAL JOURNAL OF ENGINEERING SCIENCES & RESEARCH TECHNOLOGY FEATURE EXTRACTION AND CLASSIFICATION OF TWO-CLASS MOTOR IMAGERY BASED BRAIN COMPUTER INTERFACES Geetika Kaushik and Prof Ashok Kajla Arya Institute of Engineering and Technology, Jaipur, Rajasthan, India ABSTRACT Brain-computer interface (BCI) technology provides a means of communication for people with severe movement disability to communicate with the external world using the electroencephalogram (EEG). In this study, we propose a novel method for extracting the power information contained in specific frequency bands in the context of EEG based BCI. In a two-class Motor Imagery (MI) based BCI, the main objective is to filter EEG signals before feature extraction and classification to achieve higher classification accuracy. First the EEG signal is band-pass filtered in the range of 7Hz and 25 Hz and then the mu and beta rhythms are extracted as features. LDA is then applied as the classifier for classification of left and right motor imagery. It should be noted that the proposed method improves BCI performance when compared to the raw EEG signal. KEYWORDS: Brain Computer Interface (BCI), Electroencephalography (EEG), Classification methods, BCI Research, Feature Extraction. INTRODUCTION A Brain Computer Interface (BCI) can be defined as a system which enables the user to act on his/her environment using their brainwaves without using the brain s normal output pathways which involves the peripheral nerves and muscles [1], [2]. The BCIs were designed with the initial intention of providing the ability to communicate to the handicapped population [2], [3]. The challenge in Motor Imagery-based BCI (MI-BCI), which translates the mental imagination of movement to commands, is the huge inter-subject variability with respect to the characteristics of the brain signals [4]. It backs up the belief of [5] claiming that the intention of BCIs was to enable people to converse with and control devices in the outside world without the use of traditional neuromuscular pathways thus bringing hope to those suffering from severe motor disabilities[6]. BCIs can also be used to operate devices such as prostheses enabling the paralyzed person to obtain direct brain control of a paretic limb, thus opening the way for rehabilitative and assistive applications [7]. It has been claimed that through certain mental actions our EEG can be regulated in such a way so as to operate a brain computer interface (BCI) which will change the EEG patterns into commands to allow the individual to operate devices such as prostheses [8]. However, it allows signals to be taken from the scalp and as they are non-invasive they are considered to be the safest and simplest way of recording the electrical activity of the brain [2]. As the signals recorded on EEG are recorded on the scalp they are more vulnerable to artifacts and power line interference that may be generated from muscular contractions, optical movements and other outside sources [9]. The removal of these artifacts or distortions from EEG signals before extracting features for classification of MIs is a useful step in order to increase signal-tonoise ratio (SNR) and also to obtain better separation of features of EEG signals corresponding to different imagined tasks [8]. One major challenge of this field is thus to extract reliable information from noisy data in real time in the form of relevant features [9]. These can then be passed on to classification techniques for identifying the user s mental state. Physiological arguments suggest that the µ and β frequency bands (around 8-12 Hz and 16-24 Hz) are especially relevant for discriminating motor activity. A common approach in the BCI field is thus to extract the power information from the signal over these frequency bands and use that as the classification feature. In this paper, we propose a novel use of band-pass filtering method and then computing the band power feature information contained in different frequency bands of a real signal, in the specific context of motor imagery (i.e., [746]

[Kaushik, 4(8): Augusts, 215] ISSN: 2277-9655 limb movement imagination) for Brain-Computer Interfaces (BCI) [1]. The specificity of the BCI field is to deal with very noisy data captured by Electroencephalography (EEG). The goal is to infer the user s mental state from his/her EEG signals. In a BCI setup, the task is to find whether the user imagines left hand or right hand movements, in order to build an interactive Human-Computer interface based only on brain activity. Practical applications are envisioned for severely disabled people who cannot move their limbs but whose brain is still functional [2], [9]. MATERIALS AND METHODS In this paper, we propose a use of band-pass filter between range of 7 Hz to 25 Hz to obtain enhanced EEG signals corresponding to the mu (μ) and beta (β) rhythms. The band power features are then computed from the enhanced EEG signals. These features are then classified into left and right hand MIs using an LDA classifier. The block diagram of the proposed method is shown in Fig. 1. Segmented EEG Signals Bandpass filtering between 7 Hz 25 Hz Enhanced EEG signal corresponding to mu and beta rhythms Bandpower Mu Band Feature computation Bandpower Beta Band Feature computation Bandpower Mu and beta Band Features computation LDA classifier Left hand Imagery Right hand Imagery Fig. 1. Block diagram of the proposed method [747]

[Kaushik, 4(8): Augusts, 215] ISSN: 2277-9655 Raw EEG Signal In the below diagrams we show the raw EEG data and extracted enhanced EEG signal from the raw signal after applying band-pass filtering for the complete 8 second data with sampling frequency of 25 Hz. 3 2 1-1 -2 5 1 15 2 Power 1*log 1 ( V 2 /Hz) 5-5 -1-15 -2-25 2 4 6 8 1 Frequency 1 5-5 Band Pass Filtered between 7 Hz - 25 Hz -1 5 1 15 2 Power 1*log 1 ( V 2 /Hz) -1-2 -3-4 -5-6 2 4 6 8 1 Frequency 15 1 5-5 Raw EEG Signal -1 5 1 15 2 1 5-5 Band Pass Filtered between 7 Hz - 25 Hz 2 4 6 8 1 Frequency -6-1 5 1 15 2 2 4 6 8 1 Frequency Fig. 2. The raw EEG data in trial #1 of channel C3 left hand and right hand Motor Imagery data and corresponding computed power in frequency domain and also after applying band-pass filtering between 7Hz to 25 Hz, and corresponding computed power in frequency domain. FEATURE EXTRACTION The process of feature extraction involves the extraction of signal features which will encode the intent of the user of the system [2]. As [9] explain in order to have an effective BCI operation the EEG features extracted should have a [748] Power 1*log 1 ( V 2 /Hz) Power 1*log 1 ( V 2 /Hz) 5-5 -1-15 -2-25 -1-2 -3-4 -5

[Kaushik, 4(8): Augusts, 215] ISSN: 2277-9655 strong relationship to the intent of the user. The aim of feature extraction is to transform raw brain signals into a format which makes classification easier so that noise and other unwanted, unnecessary information is removed from the input signals while still retaining the information required to differentiate between different classes of signals [11], [12]. The main problem in all brain computer interfaces is the extraction of all signals which are of interest in the presence of a large number of biological and externally induced artifacts [12]. As [2] explain brain signals can become contaminated with electrical noise during the integration of BCI with robotics or functional electrical stimulation (FES). The possibility of artificial neural contamination is also possible if activities such as muscle movements or the movement of limbs are responsible for the activation of areas which are routinely used as sources of information for the BCI [7]. Band Power Features Band Power is an estimation of the power information that is contained within different frequency bands of the EEG signals in a BCI interface. We have studied power extracted from three frequency bands known as mu feature, beta feature and mu + beta feature collectively. Feature mu is based on the EEG Mu rhythm which operates between 8-12 Hz and feature beta refers to the EEG beta rhythm which operates at a higher frequency of 16-24 Hz. Feature mu+beta uses a combination of frequency range 8-12 Hz and 16-24 Hz [11]. LINEAR DISCRIMINANT ANALYSIS (LDA) CLASSIFIER It is a robust and linear classifier operating under the assumption that the different classes can be separated linearly. It gives a hyper plane to distinguish the features into two different classes. In general linear classifiers are more robust than their non-linear classifiers as they have limited flexibility and there is less chance of over fitting [13]. Generally, it is not straightforward to classify the extracted features for classification of EEG signals in BCI applications. Finding the optimum combination of the features which can provide better discrimination, is a crucial task. In this paper, we have used LDA classifier which is commonly used in EEG-based BCI applications. The LDA based classifier tries to reduce the dimensionality and at the same time it protects most of the class discriminatory information. Suppose, we have a set of two classes denoted by w 1, w 2. Then, we classify the n-dimensional sample points x = {x1, x2, x3, xn}, n 1 samples to the class w 1, and n 2 samples to the class w 2. In this method, we try to achieve a line y = w t x from the set of all possible lines, and that line maximizes the discrimination between the two considered classes. In order to obtain a good projection vector, we require to measure the separation between the two classes chosen for the study. The mean vector of each class in x-space and y-space is given by following equations: µ i = 1 N i x xεw i, (1) andʉ i = 1 y = 1 w t x = w t µ N i yεw i N i (2) i yεw i The objective function is defined as the distance between the two projected means. It can be expressed as follows: J (w) = ʉ 1 ʉ 2 = w t (µ 1 µ 2 ) (3) However, the distance measured between projected means may not always be a good measure because the standard deviation between classes has not been considered. In order to overcome the limitation stated above, the [749]

[Kaushik, 4(8): Augusts, 215] ISSN: 2277-9655 enhancement of LDA has been proposed which is known as Fishers LDA classifier. It determines a decision boundary or probably a hyperplane in the feature space to classify the features in to distinct classes. It finds out the separation boundary between two given distributions in terms of ratio of two group variances as given below [3]: J(w) = σ 2 between 2 = wt (µ 1 µ 2 ) 2 σ within w t S 1 w + w t S 2 w where µ 1, µ 2 are the mean of the classes and S 1, S 2 are the variances of the feature distributions between two classes w1, w2 respectively. The maximum separation between two classes can be shown by (1) as: w = (S 1 + S 2 ) 1 (µ 1 µ 2 ) (5) The w is weight vector which provides optimum direction of projection of the data. In the Fisher s LDA, the decision boundary uses the following equation to classify the feature vector d(m) as [14]: p(m) = d(m)w t + b (6) where b is the threshold or bias. The features are assigned to one of the classes based on the sign of the p(m). (4) RESULTS AND DISCUSSION In this work, the BCI competition IV dataset 2B [15] has been used to measure the performance of the proposed method. The dataset contains EEG signals of two classes of three bipolar channels from 9 healthy subjects, denoted by B1-B9. In this dataset, for each subject 5 sessions has been provided. The data recorded from three bipolar channels include C3, Cz and C4 with a sampling frequency of 25 Hz [25]. In this paper, for training phase, a single session namely ^3T has been used. For the evaluation phase, we have used two sessions namely, ^4E and ^5E for computing the accuracy in the classification of left and right MI EEG signals. It should be noted that, the ^ in the session name denotes the subject number which ranges from B1 to B9. During the training session, we have performed a 5-fold cross validation in order to determine the time window for the best possible classification accuracy by the LDA classifier for classification of the left and right hand MI EEG signals. In our work we have considered two frequency-bands namely 8-12Hz and 16-24Hz corresponding to mu and beta bands. Hence band power features are computed in the selected frequency bands individually and collectively. The extracted features have been given as an input feature vector to the LDA classifier for classification of left and right hand MI EEG signals. Table I depicts the maximum classification accuracy obtained for BCI competition IV dataset 2B with band-pass filtering, which provides the enhanced EEG signals between 7 Hz and 25 Hz for the nine subjects denoted by B1- B9 across three sessions 3T, 4E, and 5E. Table I MAXIMUM CLASSIFICATION ACCURACIES OBTAINED IN MU BAND ON BCI COMPETITION IV DATASET 2B. Subject Accuracy with Band-pass filtering Mu(8 Hz -12 Hz) (%) Channel Selected Training Evaluation (3T) (4E) (5E) B1 8.63 7 71.88 C3 C4 Cz B2 8.63 79.17 65 C3 C4 Cz B3 86.25 9 86.88 C3 C4 Cz B4 81.88 8 8 C3 C4 Cz [75]

[Kaushik, 4(8): Augusts, 215] ISSN: 2277-9655 B5 71.25 53.75 81.25 C3 C4 B6 71.88 71.88 64.38 C3 C4 Cz B7 65.63 61.88 54.38 C3 C4 Cz B8 88.75 93.75 93.13 C3 C4 Cz B9 7.63 61.25 61.25 C4 Cz Average 77.5 73.52 73.13 Std 7.91 13.45 12.94 As mentioned earlier, we have considered channels C3, C4 and Cz for calculating the results. In the training stage, after applying band-pass filtering, the average accuracy obtained across the nine subjects is 77.5 % with standard deviation of 7.91. In the evaluation stages, the average accuracies obtained in sessions 4 and 5 are 73.52 % and 73.13% respectively. Also, the standard deviations obtained are 13.45 and 12.94 respectively. Table II MAXIMUM CLASSIFICATION ACCURACIES OBTAINED IN BETA BAND ON BCI COMPETITION IV DATASET 2B. Subject Accuracy with Band-pass filtering Beta(16 Hz -24 Hz) (%) Channel Selected Training Evaluation (3T) (4E) (5E) B1 65.63 51.88 63.75 C4 Cz B2 71.25 72.5 55 C3 C4 B3 93.13 9 86.88 C3 C4 Cz B4 84.38 75.63 72.5 C3 C4 Cz B5 59.38 5.63 64.38 C4 B6 56.88 68.13 59.38 C3 C4 Cz B7 89.38 77.5 8.63 C3 C4 Cz B8 86.88 84.38 76.25 C3 Cz B9 88.75 88.75 93.75 C3 C4 Cz Average 77.29 73.26 72.5 Std 14.6 14.42 13.4 As mentioned earlier, we have considered channels C3, C4 and Cz for calculating the results. In the training stage, after applying band-pass filtering, the average accuracy obtained across the nine subjects is 77.5 % with standard [751]

[Kaushik, 4(8): Augusts, 215] ISSN: 2277-9655 deviation of 7.91. In the evaluation stages, the average accuracies obtained in sessions 4 and 5 are 73.52 % and 73.13% respectively. Also, the standard deviations obtained are 13.45 and 12.94 respectively. Table II shows the maximum classification accuracy obtained with band-pass filtering in the range of 7 Hz and 25 Hz for the nine subjects denoted by B1-B9 across three sessions 3T, 4E, and 5E for BCI competition IV dataset 2B. In the training stage, the average accuracy calculated across the nine subjects is 77.29 % with standard deviation of 14.6 and the average accuracy in evaluation session 4E is 73.6 % with standard deviation of 14.42 and in the evaluation session 5E is 72.5% with standard deviation of 13.4. Table III depicts the maximum classification accuracy obtained for BCI competition IV dataset 2B for the enhanced EEG signals obtained with band-pass filtering in the range of 7 Hz and 25 Hz for the nine subjects denoted by B1- B9 across three sessions 3T, 4E, and 5E. Table III MAXIMUM CLASSIFICATION ACCURACIES OBTAINED IN MU AND BETA BAND ON BCI COMPETITION IV DATASET 2B. Subject Accuracy with Band-pass filtering Mu(8 Hz -12 Hz) and beta (16 Hz-24 Hz) Collectively(%) Channel Selected Training Evaluation (3T) (4E) (5E) B1 81.88 75 71.25 C3 C4 Cz B2 8 72.5 68.75 C3 C4 Cz B3 95.63 92.5 87.5 C3 C4 Cz B4 85.63 85 79.38 C3 C4 Cz B5 79.38 62.5 64.38 C3 C4 Cz B6 76.25 71.88 65.63 C3 C4 Cz B7 89.38 76.88 81.25 C3 Cz B8 9.63 93.13 92.5 Cz B9 91.25 86.88 89.38 C3 Cz Average 85.56 79.58 77.78 Std 6.55 1.38 1.68 In the training stage, after applying band-pass filtering, the average accuracy obtained across the nine subjects is 85.56 % with standard deviation of 6.55. In the evaluation stages, the average accuracies obtained in sessions 4 and 5 are 79.58 % and 77.78% respectively. Also, the standard deviations obtained are 1.38 and 1.68 respectively. The mu and beta bands feature collectively show higher classification accuracies with band-pass filtering, there is significant improvement of approximately 8% (p<.1) of the maximum classification accuracy for all subjects across the training session 3T. In the evaluation sessions, with the feature combination of mu and beta bands the average classification accuracy improvement is > 6 % (p <.1) for 4E and in 5E, the average classification accuracy improvement is > 4 % (p <.1). Fig. 3 displays the maximum classification accuracy obtained across all the nine subjects for training session 3T. The performance improvements across all the nine subjects using a combination of mu and beta band power features [752]

[Kaushik, 4(8): Augusts, 215] ISSN: 2277-9655 are illustrated with the bar graphs. In Fig. 4 and Fig. 5, the combination of mu and beta band power features shows an improvement in classification accuracy. In training session 3T, eight out of the nine subjects have shown significant improvement. On the other hand, in evaluation session 4E, five out of the nine subjects have shown improvement in the classification accuracy as compared to mu feature and beta feature independently and in 5E, with a combination of mu and beta, and improvement in classification accuracy (p<.1). 1 9 Training session (3T) Mu Beta Mu + Beta 8 Classification accuracy (%) 7 6 5 4 3 2 1 1 2 3 4 5 6 7 8 9 Subjects Fig. 3. Accuracy difference across all the features. The graph show an improvement of classification accuracy when we use collectively use mu + beta features in comparison to individual mu and beta features across training session (3T). 1 Evaluation session (4E) 9 8 Mu Beta Mu + Beta Classification accuracy (%) 7 6 5 4 3 2 1 1 2 3 4 5 6 7 8 9 Subjects Fig. 4. The graph show an improvement of classification accuracy when we use collectively use mu + beta features in [753]

[Kaushik, 4(8): Augusts, 215] ISSN: 2277-9655 1 9 comparison to individual mu and beta features across training session (4E). Evaluation session (5E) Mu Beta Mu + Beta 8 Classification accuracy (%) 7 6 5 4 3 2 1 1 2 3 4 5 6 7 8 9 Subjects Fig. 5. The graph shows an improvement of classification accuracy when we use collectively use mu + beta features in comparison to individual mu and beta features across training session (5E). CONCLUSION We have applied band-pass filtering to extract the meaningful EEG data in order to correctly classify the left hand and right hand motor imagery task before feature extraction step. The filtering method help us to reduce the effect of artifacts and to enhance the performance of a motor imagery based brain-computer interface (BCI). A combination of features whose frequencies fall in the frequency range of μ and β rhythms has provided improvement in the accuracy of classifying left and right hand MI EEG signals as compared to μ and β feature individually. In future, it would be of interest to develop and implement new features along with the band-pass filtering method for classifying MI EEG signals. Although enhanced feature separability offered by the band-pass filtering method has helped to increase the classification accuracy by reducing the effect of artifacts. REFERENCES 1. J. R. Wolpaw, N. Birbaumer, W. J. Heetderks, D. J. McFarland, P. H. Peckham, G. Schalk, E. Donchin, L. a Quatrano, C. J. Robinson, and T. M. Vaughan, Brain-computer interface technology: a review of the first international meeting., IEEE Trans. Rehabil. Eng., vol. 8, no. 2, pp. 164 73, Jun. 2. 2. J. R. Wolpaw, N. Birbaumer, D. J. McFarland, G. Pfurtscheller, and T. M. Vaughan, Brain-computer interfaces for communication and control., Clin. Neurophysiol., vol. 113, no. 6, pp. 767 91, Jun. 22. 3. B. Fabio, Brain Computer Interfaces for communication and control, Front. Neurosci., vol. 4, pp. 767 791, 19. 4. S. Lemm, B. Blankertz, T. Dickhaus, and K. R. Müller, Introduction to machine learning for brain imaging, Neuroimage, vol. 56, no. 2, pp. 387 399, 211. 5. R. N. Roy, S. Bonnet, S. Charbonnier, and A. Campagne, Mental fatigue and working memory load estimation: Interaction and implications for EEG-based passive BCI, Proc. Annu. Int. Conf. IEEE Eng. Med. Biol. Soc. EMBS, pp. 667 661, 213. 6. S. Shahid, R. Sinha, and G. Prasad, Mu and beta rhythm modulations in motor imagery related post-stroke EEG: a study under BCI framework for post-stroke rehabilitation, in BMC Neuroscience, 21, vol. 11, no. Suppl 1, p. P127. 7. G. Prasad, P. Herman, D. Coyle, S. McDonough, and J. Crosbie, Applying a brain-computer interface to support motor imagery practice in people with stroke for upper limb recovery: a feasibility study., J. Neuroeng. Rehabil., vol. 7, no. 1, p. 6, 21. 8. P. Gaur and H. Wang, An Empirical Mode Decomposition Based Filtering Method for Classification of Motor-Imagery EEG Signals for Enhancing Brain-Computer Interface. 9. J. N. Mak and J. R. Wolpaw, Clinical Applications of Brain-Computer Interfaces: Current State and Future Prospects., IEEE Rev. Biomed. Eng., vol. 2, pp. 187 199, 29. [754]

[Kaushik, 4(8): Augusts, 215] ISSN: 2277-9655 1. P. L. Lee, J. J. Sie, Y. J. Liu, C. H. Wu, M. H. Lee, C. H. Shu, P. H. Li, C. W. Sun, and K. K. Shyu, An SSVEP-actuated brain computer interface using phase-tagged flickering sequences: A cursor system, Ann. Biomed. Eng., vol. 38, no. 7, pp. 2383 2397, 21. 11. P. Herman, G. Prasad, T. M. McGinnity, and D. Coyle, Comparative Analysis of Spectral Approaches to Feature Extraction for EEG-Bdased Motor Imagery Classification, IEEE Trans. Neural Syst. Rehabil. Eng., vol. 16, no. 4, pp. 317 26, 28. 12. S. Sun and J. Zhou, A Review of Adaptive Feature Extraction and Classification Methods for EEG-Based Brain-Computer Interfaces, 214. 13. F. Lotte, M. Congedo, L. Anatole, F. Lamarche, and B. A. A, A review of classification algorithms for EEG-based brain computer interfaces To cite this version : A Review of Classification Algorithms for EEG-based Brain-Computer Interfaces, J. Neural Eng., 27. 14. A. F. Marquand, S. De Simoni, O. G. O Daly, S. C. R. Williams, J. Mourão-Miranda, and M. A. Mehta, Pattern classification of working memory networks reveals differential effects of methylphenidate, atomoxetine, and placebo in healthy volunteers., Neuropsychopharmacology, vol. 36, no. 6, pp. 1237 47, May 211. 15. C. Brunner, R. Leeb, and C. Brunner, BCI Competition 28 Graz data set B, pp. 1 6, 28. [755]