I'm trying to compute a dissimilarity matrix based on a big data frame with both numerical and categorical features. When I run the daisy
function from the cluster package I get the error message:
Error: cannot allocate vector of size X.
In my case X is about 800 GB. Any idea how I can deal with this problem? Additionally it would be also great if someone could help me to run the function in parallel cores. Below you can find the function that computes the dissimilarity matrix on the iris dataset:
require(cluster)
d <- daisy(iris)
I've had a similar issue before. Running
daisy()
on even 5k rows of my dataset took a really long time.I ended up using the
kmeans
algorithm in theh2o
package which parallelizes and 1-hot encodes categorical data. I would just make sure to center and scale your data (mean 0 w/ stdev = 1) before plugging it intoh2o.kmeans
. This is so that the clustering algorithm doesn't prioritize columns that have high nominal differences (since it's trying to minimize the distance calculation). I used thescale()
function.After installing h2o: