Hi All,
I am currently trying to familiarize with "ff" package which allows me
to store R objects in the hard drive. One of things that I notice when reading
in a text file with "read.csv.ffdf" function - is that, in the R temp
folder, 1000+ small files get created, each file having a name like
"ffdf1bcd8b4aa0.ff". Each file is about 5KB in size.
My understanding is, the whole file has been split into small small pieces and
stored in the hard drive. What I am trying to see is that - is there a way to
reduce the number of splits by increasing the file size?
Thanks in advance,
Regards,
Indrajit
[[alternative HTML version deleted]]