similar to: File too big for filehash?

Displaying 20 results from an estimated 100 matches similar to: "File too big for filehash?"

2010 Apr 09
5
Read data in sequences
Dear R users, I tried to find a solution in the search list, but I cannot find it. I would like to read a .txt file with, let say, three variables, with two of which have repeated values in a number a columns. An example: The variables: Treat, x1, x2. The values: A 2.5 3.4 2.7 5.6 5.7 5.4 10.1 9.4 ... B 5.3 5.4 6.5 7.5 1.3 4.5 10.5 4.1 ... ... In the first column, the letters represent the
2009 Jan 23
0
Appending objects created using filehash package
Hi, I am working with a very large dataset, and am using the 'filehash' package to manage such a large file. While I have no problem accessing objects that I load into a database, I was hoping there is a better way to append to objects already in the database. The only way I know now to append to an object, basically requires rewriting the entire object. Sample code:
2009 Oct 01
2
Huge matrix: allocation works but assignment fails
Hello everyone, I am working with one big matrix: w=matrix(0,18000,18000) on a Linux computer with 16Go of RAM. I can actually create the matrix, and even access elements: w[10,10] 0 but if I try to change one element, it fails: w[10,10]=1 Erreur : impossible d'allouer un vecteur de taille 2531250 Ko (Failed to allocate a vector of size...) What can I do? And, maybe even more important,
2012 Aug 29
1
spatial correlation in lme and huge correlation matrix (memory limit)
Hi, I'm trying to introduce a (spatial) exponential correlation structure (with range=200 and nugget.effet of 0.3) in a lme model of this form: lme(ARBUS~YEAR, random=~1|IDSOUS). The structure of the data is "IDSOUS" "XMIN" "YMAX" "YEAR" "ARBUS" with 2 years of data and 5600 points for each year. I do:
2011 Apr 13
2
Dump the "source code" of data frame
Dear R experts, I remember a similar function existed and have been mentioned in R-help before. I tried my best to search but I really can't find it out. suppose I have an data frame like this: > somedata <- data.frame(age.min = 1, age.max = 1.5, male = TRUE, l = -1.013, m=16.133, s=0.07656) In order to back up the data and I don't want to use write.table(), I would like to back
2011 Aug 02
2
Memory limit in Aggregate()
Dear all, I am trying to aggregate a table (divided in two lists here), but get a memory error. Here is the code I'm running : sessionInfo() print(paste("memory.limit() ", memory.limit())) print(paste("memory.size() ", memory.size())) print(paste("memory.size(TRUE) ", memory.size(TRUE))) print(paste("size listX ", object.size(listX)))
2007 Jul 04
0
Kmeans performance difference
Hi All, A question from a newbie using R 2-5-0 on windows XP. Why is it that kmeans clustering with apparently the exact same parameters behaves so differently between the two following examples : > cl1 <- kmeans(subset(pointsUXO15555, select = c(2:4)), 10) Takes about 2 seconds to deliver a result > cl1 <- clust(subset(pointsUXO15555, select = c(2:4)), k=10,
2012 Mar 01
0
Memory problem in R
Hi all, I am running an -MNP- multinomial probit model package using R. It gives me the following objection instead of giving me the results: Erreur : impossible d'allouer un vecteur de taille 137.9 Mo (in english: cannot allocate a 137.9 Mb vector memory).  I have already increased the memory size upto 2047Mb. This problem has been discussed in 2008 (archives) but no profitable answers were
2011 Mar 10
0
Avoiding choosing parameters with mix[mixdist]
Hi, I am working on a population of an invasive clam. The data are the size of each clam per station (2mm on average). Each station is found at a different distance from a power nuclear station, so at different water temperatures. The fist step I want to do is to identify cohort size at each station or (zone of water temperature). The second step will be to see whether the size or number of
2012 Jul 23
2
Large data set
Hi all, Have a problem. Trying to read in a data set that has about 112,000,000 rows and 8 columns and obviously enough it was too big for R to handle. The columns are mode up of 2 integer columns and 6 logical columns. The text file is about 4.2 Gb in size. Also I have 4 Gb of RAM and 218 Gb of available space on the hard drive. I tried the dumpDF function but it was too big. Also tried bring in
2012 May 04
2
Can't import this 4GB DATASET
Dear Experienced R Practitioners, I have 4GB .txt data called "dataset.txt" and have attempted to use *ff, bigmemory, filehash and sqldf *packages to import it, but have had no success. The readLines output of this data is: readLines("dataset.txt",n=20) [1] " "
2008 Aug 28
0
Can the file locking in filehash be reused? (Was: Re: [R] [R-pkgs] filehash 2.0)
Hi (Roger), I saw the announcement of filehash v2.0 and the sentence "This development has lead to better file locking for concurrent access and faster reading and writing of data in general" caught my attention. What kind of file locking do you refer to here? I am looking for a mechanism that can be used to lock files for reading and/or writing, and I'd love to have a cross
2008 Aug 28
0
filehash 2.0
I have just uploaded to CRAN version 2.0 of the 'filehash' package. This version contains a major rewriting of many of the internals (much rewritten in C) for the DB1 format, which is the default. This development has lead to better file locking for concurrent access and faster reading and writing of data in general. In addition to rewriting the internals, I have added two modules for a
2008 Aug 28
0
filehash 2.0
I have just uploaded to CRAN version 2.0 of the 'filehash' package. This version contains a major rewriting of many of the internals (much rewritten in C) for the DB1 format, which is the default. This development has lead to better file locking for concurrent access and faster reading and writing of data in general. In addition to rewriting the internals, I have added two modules for a
2010 Jan 02
0
filehash - multiple indices via '[' not allowed when using RDS format
Hi, I have been using filehash for a while. It has performed very well. However, recently I found filehash gives an error when I need to do something like db[c("a", "b")] when the db is in RDS format. Does any one know a way to get around that? The code below reproduces the error thanks Jeff filehashOption(defaultType = "DB1") dbCreate("mydb3", type =
2008 Mar 15
1
filehash
Hello, I'm using filehash on the windows XP and it has been working fine with the newest R version 2.6.2. However, on the windows vista, when I ran the same code, I got the following error: > dbCreate("simdb") #create simdb database [1] TRUE > db<-dbInit("simdb") #initiate an object of database Error in sprintf(gettext(fmt, domain = domain), ...) : object
2010 Jan 21
0
filehash does not install on FreeBSD
Trying to install package 'filehash' I get the following error on FreeBSD 9.0-CURRENT (amd64) with R version 2.11.0 (2010-01-15 r50990): ----------------------------------- R CMD INSTALL filehash_2.0-1.tar.gz * installing to library '/usr/local/lib/R/library' * installing *source* package 'filehash' ... ** libs gcc -std=gnu99 -I/usr/local/lib/R/include
2011 Jan 02
1
filehash for big data
Hi all, I am trying to use the filehash library to analyze a 5M by 20 matrix with both double and string data types. After consulting a few tutorials online, it seems as though one needs to first read the data into R; then create an R object; and then assign that object a location in my computer via filehash. It seems like the benefit of this is minimizing memory allocation when running
2010 Feb 22
1
big panel: filehash, bigmemory or other
Dear R-list I'm on my way to start a new project on a rather big panel, consisting of approximately 8 million observations in 30 waves of data and about 15 variables. I have a similar data set that is approximately 7 gigabytes in size. Until now I have done my data management in SAS, and Stata, mostly identifying spells, counting events in intervals, and a like, but I would like to
2001 Nov 05
1
Problem to transfer Splus functions
Hello I would like to transfer some Splus functions in R. But I have a problem first about this assignation in Splus : xnom <- deparse(substitute(x)) I am a bad programmer : I don't understand the R help How to modify these functions ? Thank you very much for your help Here are the four functions and a data test