-
Notifications
You must be signed in to change notification settings - Fork 21
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Overfitting when training lstm model! #1
Comments
Hi |
Hi, the results in the paper are affected by slow frequency in EEG data. When filtering properly the data, the results for lstm are about 15-20% as you obtained. |
Thanks for your response. |
We cannot release ImageNet data for licence issues. What we can do is to share the class names. Sorry. |
Thanks for your response. |
If you run the code on the raw data you will get 80% of accuracy, if you filter the data from 1Hz to 45Hz you obtain the 20% of accuracy. |
Thanks a lot |
@cspampin So, is the accuracy reported in the paper for CVPR and TPAMI using raw eeg signals? |
@YutLan What I observe is results are reproducible with raw file data but not with 5_95 hz filter one. I have cross-checked with all other methods in different papers and reached the same conclusion. We tried reproducing it with 5_95 Hz for a long time but when I saw this repo then I realise that it is possible with Raw data. Now only thing I am not sure about is the results of the PAMI paper. It will be helpful if you crosscheck this from your end. |
@prajwalsingh Thank you very much! I have downloaded the dataset. Does the raw data mean the file "eeg_signals_raw_with_mean_std.pth"? |
@YutLan , Yes. |
@prajwalsingh Thank you! It will be helpfu! Maybe my implementation misused the early stop. I will crosscheck this from my end! |
By the way, Nice to meet you! @prajwalsingh I've actually read your team's eeg2image paper before. What a coincidence! |
Thank you 😊, good to know that. |
I have searched the learning rate, seed with the raw data. The best results on the valid and test dataset are about 40%. Would you so kind to tell me your experimental details? @prajwalsingh |
@YutLan You can try this code. I have used this got good ACC: |
Thank you! I will reply as soon as possible! |
@prajwalsingh TrL=0.7866, TrA=0.7243, VL=1.6355, VA=0.5029, TeL=1.6905, TeA=0.5020, TeA at max VA = 0.5312 at epoch 2045, it seems that it not work and maybe I should increase the epochs. Would you be so kind as to share the data link you used? |
@YutLan I have used the same dataset. I downloaded the code from this repository only. Try changing lstm depth from 1 to 2 or 4. You may get acc around 94% on test data. |
@prajwalsingh Thanks, It works! |
I just visualize some EEG signals belong to the same class, using
subj 4 & label 10
subj 4 & label 15
|
Is that normal? |
I also visualize some EEG signals belong to the same class of
subj 4 & label 10
|
Many thanks!!! :) |
Yeah, I found it works on raw data but not preprocessed data either. And I wonder why dose this phenomenon happen, I mean, the raw data looks too prefect (than other eeg dataset such as THINGS-eeg) isn't it? |
In the PAMI paper, they address the problem with raw data that leads to overfitting (might be some noise/signal pattern is easy to identify be neural network) as well as there are some overlaps in the train and test data. Therefore authors proposed a method for processing filter 5-95Hz in the PAMI paper. You can preprocess the raw data from your end with a new train-val-test split without overlapping samples. You can also check one of our works recently accepted in WACV24: Learning Robust Deep Visual Representations from EEG Brain Recordings [https://arxiv.org/abs/2310.16532] |
Oh! Thank you for your guidance. I realized I was reading an older version of the paper. Does the new version propose corrections to the original experiments? I will read it to understand. I will also review your work. If there are areas I don't understand, I hope to seek your guidance. :) |
I find something interesting 🤣 . In another TPAMI paper The Perils and Pitfalls of Block Design for EEG Classification Experiments, it said that "We take our correspondence with the authors to imply that no filtering was applied during acquisition, no filtering was applied prior to production of either the Python or Matlab format released data". And I have found some more papers which think this dataset has some flaws. Maybe I should change the dataset 😢 |
Interesting, thanks for sharing the details. You can check the ThoughtViz paper and one more object EEG dataset is there. I think there are few papers in CVPR21 that proposed the EEG dataset. |
Hi, could you please share the code again? The link is not available. Many thanks! |
Hi, I have updated the [link]. |
Wow, thanks for your quick reply, I will try this code! |
Hello, could you please share your opinion about dataset splitting? I noticed that the paper mentioned, "Our dataset was split into training, validation, and test sets, with respective fractions 80% (1600 images), 10% (200 images), 10% (200 images)." Does this mean that the final size of the training set is 16006=9600, the validation set is 2006=1200, and the test set is 200*6=1200? However, in the file block_splits_by_image_single.pth, the lengths of the three are 669:167:164, and in block_splits_by_image_all.pth, the lengths are 7970:1998:1997. These are not in an 8:1:1 ratio. What is the reason for this? Many thanks! |
Hi YutLan, I read your "Seeing through the Brain" paper before. I'm trying to download the dataset from OneDrive now, but I'm finding that the download is unusually slow in mainland China, and I've been unable to get it to download successfully. Could you please share a download link with me privately? Thank you very much! |
ref: perceivelab#1 Signed-off-by: Tariq Mehmood <[email protected]>
ref: perceivelab#1 (comment) Signed-off-by: Tariq Mehmood <[email protected]>
Hi @cspampin, thanks for your great work firstly. Currently, I've trained the model (lstm) in your dataset. But after training for 100 epochs, the model was overfitting, TrL was decreasing while VL and TeL were increasing, and TeA at max VA = 0.1457 at epoch 28. I wonder how can I fix the problem to achieve good performance (as result in the paper).
The text was updated successfully, but these errors were encountered: