SJP SCI-TECH CLUB
  • Home
  • Student Research
    • Eel Grass Studies
    • Aquaponics Blog
    • Wind Energy Research
    • Deep Learning for BCI
    • Cloud Chamber Blog
    • And much more.. >
      • Bioluminesence
  • Lab Visits
    • Novartis Cambridge
    • Greentown Labs
    • MASS CEC
    • MIT Plasma Physics Center
    • Histogenics
    • US GreenBuild - Boston
  • Physics Olympics
    • Paper Airplanes
    • Glider Competition
  • Internships + More
    • Histogenics 2017
Picture

Topic for the future

4/23/2018

 
Picture
Because different people have different patterns of brain signal, it is extreme difficult or nearly impossible to use the model trained by one person's brain activity for another person. However, the technique called transfer learning can make the algorithm remain stable. There were some researchers in the world applying the transfer learning to classical algorithms, such as SVM, but few people apply this technique to deep learning algorithms in the field of Brain Computer Interface. Designing a deep transfer learning algorithm can be the next step for my deep learning research. 

Result and Conclusion

4/23/2018

 

Test Accuracy for the AE-LSTM-FFN Algorithm

Picture

Test Accuracy for the Support Vector Machine algorithm

Picture
The best version of my algorithm achieves about 75% accuracy. The classical Support Vector Machine algorithm (SVM), however, has only 70% accuracy in average. In addition to the accuracy that my algorithm can achieve, it also has smaller variance, compared to the SVM. SVM algorithm has a variance around 20% in average, but my algorithm has only 5% variance due to the effectiveness of the autoencoder. In conclusion, the deep learning algorithm does outperform the classical algorithm. 

Dataset

4/23/2018

 
2.1. Dataset:
The data comes from BCI Competition IV (2008), datasets 2B . The dataset contains the brain activities of 9 different subjects. As the dataset description shown, “Three bipolar recordings (C3, Cz, and C4) were recorded with a sampling frequency of 250Hz. The recordings had a dynamic range of ±100µV for the screening and ±50µV for the feedback sessions. They were bandpassfiltered between 0.5Hz and 100Hz, and a notch filter at 50Hz was enabled. The placement of the three bipolar recordings (large or small distances, more anterior or posterior) were slightly different for each subject. The electrode position Fz served as EEG ground”. The recorded brain signals are then stored in the General Data Format. The GDF files can be loaded to Matlab using the BioSig toolbox.
Picture

A new approach

4/23/2018

 
An AE-LSTM-FNN architecture for the non-invasive, motor imagery Brain Computer Interface system.
1. Why Non-invasive BCI: Compare to the invasive BCI, non-invasive BCI receives brain signal with much lower quality. Although the brain signal contain more noise, but the non-invasive BCI still has advantage over invasive BCI. Non-invasive BCI is more secure than the invasive one because the infection caused by implanting the sensors can be avoided. What’s more, the neurons near the sensors inside the brain will degraded as the time goes, then the sensors becomes ineffective. However, this will not happen when the subject is using non-invasive BCI. Moreover, because of the current low accuracy of non-invasive BCI, the system can still be improved as the classification algorithm becomes more and more advanced.

​2. 
Why motor imagery: Unlike P300 and SSVEP that requires permanent attention to the external stimuli, motor imagery is independent of any stimulation and can be operated by users with free will. However, the motor imagery signal does has some disadvantages. The training costs a long time for users, may requiring several weeks or months. This is due to the brain tries to adapt Brain Computer Interface and produce more distinguishable pattern. The advanced algorithm can reduce the time of training.

Architecture of Algorithm

Picture
AutoEncoder:

The Autoencoder is an algorithm that can reduce the noise and dimensionality of the sample. This is an unsupervised training algorithm, meaning this algorithm does not require the labels for learning the patterns of the datasets. In the case of Brain Computer Interface, the autoencoder does not need the intention of moving right or left arm the users thought to learn the patterns from their brain activities. An Autoencoder consists of two parts, the encoders and decoders.  The encoders transform the original data into a compressed representation with much lower dimensions. Then, the decoders transform the compressed representation back to the representation with original dimension [10]. Autoencoders try to learn the weights of transformation that minimize the difference between original input and reconstructed inputs. Because the compressed representation has lower dimension, this encoding process is lossy. Therefore, the algorithm has to learn the weight that can preserve the information from the original input in lower dimension, resulting the compressed representation contain more useful information and less noise.
Picture
Long Short Term Memory:
Long Short Term Memory is a deep learning architecture that has the ability to learn a pattern from a time series or a sequence data. LSTM was applied to the voice recognition tasks in the past, and classifiers achieved higher accuracy than most classicial algorithms. Because the LSTM will learn the pattern of the brain activity across all the time step, it is possible that applying LSTM to classify motor imagery EEG can improve the accuracy of the algorithm. 
Picture

    Author

    Jack Lin 

    Archives

    April 2018
    March 2018

    Categories

    All

    RSS Feed

Proudly powered by Weebly
  • Home
  • Student Research
    • Eel Grass Studies
    • Aquaponics Blog
    • Wind Energy Research
    • Deep Learning for BCI
    • Cloud Chamber Blog
    • And much more.. >
      • Bioluminesence
  • Lab Visits
    • Novartis Cambridge
    • Greentown Labs
    • MASS CEC
    • MIT Plasma Physics Center
    • Histogenics
    • US GreenBuild - Boston
  • Physics Olympics
    • Paper Airplanes
    • Glider Competition
  • Internships + More
    • Histogenics 2017