Hello,
I'm using a large-scale dataset such as ML-10M or netflix and I find that the DataModel<Long,Long> object is taking up to much space, in fact I run out of memory even before loading everything into the DataModel<Long,Long> structure, I removed the timestamp variable from all samples but it didn't do the trick.
Is it me or it's expected that happen? I have 16GB of RAM so it should be more than sufficient to load a sparse matrix into memory even for the netflix "problem" which is a ~3gb trainset.
Thanks,
André