You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I found if a dataset contains cells less than 1/4 the number of cells in the other dataset, the algorithm would train without using the smaller dataset at all. Look at line 142 for batch_size assignment in scDART.py and line 151 in train.py.
Also I am curious about the way you set bacth_size (1/4 of the larger dataset), would that be unnecessarily too large for dataset contain hundreds of thousands of cells in terms of memory usage? Will that be a concern if we fix batch_size at smaller number like 128 (this is what I saw other auto-encoder based algorithms used)?
The text was updated successfully, but these errors were encountered:
Thanks for pointing out the issue. We have updated the code in line 142batch_size assignment. Instead of using the max, we use min. Yes, you can set your own batch size, and we have that option in the scDART model: batch_size parameter in class initialization. We chose a large batch size as it performs better in the tests of the manuscript.
I found if a dataset contains cells less than 1/4 the number of cells in the other dataset, the algorithm would train without using the smaller dataset at all. Look at line 142 for batch_size assignment in scDART.py and line 151 in train.py.
Also I am curious about the way you set bacth_size (1/4 of the larger dataset), would that be unnecessarily too large for dataset contain hundreds of thousands of cells in terms of memory usage? Will that be a concern if we fix batch_size at smaller number like 128 (this is what I saw other auto-encoder based algorithms used)?
The text was updated successfully, but these errors were encountered: