Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

R crash with large datasets #3

Open
rmathieu25 opened this issue Aug 3, 2021 · 2 comments
Open

R crash with large datasets #3

rmathieu25 opened this issue Aug 3, 2021 · 2 comments

Comments

@rmathieu25
Copy link

rmathieu25 commented Aug 3, 2021

Hello,

Thank you for this package that seems very useful.

Nevertheless, I have an integrated datasets with 150 000 cells, and my R crash when calculating the co-clustering frequencies. I have a computer with a RAM memory of 128 GB. My integrated dataset takes up 15.4GB. Do you have any idea how could I overcome this issue.

Thank you in advance

@MenonLab
Copy link

MenonLab commented Aug 3, 2021

Hi there,

Thanks for mentioning this - we have not yet optimized chooseR for very large data sets, but will let you know when we update the code to allow for that. In the meantime, is there any substructure at the top level of your data set? If there are clearly discrete subgroups of cells at the top level, you could run chooseR separately on these subsets - usually, the subsets are also where there is more confusion about what parameter values are the most robust.

Also, we have forked the development version of the code here:
https://github.com/MenonLab/chooseR
We hope to have the updated, more memory-efficient version soon.

@rmathieu25
Copy link
Author

Hello,

Thank you for your answer.

I was thinking to run chooseR on subsets so I will give it a try.

Thank you.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants