similar to: Large data set

Displaying 14 results from an estimated 14 matches similar to: "Large data set"

2012 Jul 04
1
RODBC tables
Dear Sir/Madam, I am desperately in need of some help. I am trying to access tables from the oracle database and inserting them into R via a data frame and I keep getting an error saying that "Error in .Call(C_RODBCFetchRows, attr(channel, "handle_ptr"), max, buffsize, : negative length vectors are not allowed". My connection is fine and my code for this is:
2011 Apr 13
2
Dump the "source code" of data frame
Dear R experts, I remember a similar function existed and have been mentioned in R-help before. I tried my best to search but I really can't find it out. suppose I have an data frame like this: > somedata <- data.frame(age.min = 1, age.max = 1.5, male = TRUE, l = -1.013, m=16.133, s=0.07656) In order to back up the data and I don't want to use write.table(), I would like to back
2012 May 04
2
Can't import this 4GB DATASET
Dear Experienced R Practitioners, I have 4GB .txt data called "dataset.txt" and have attempted to use *ff, bigmemory, filehash and sqldf *packages to import it, but have had no success. The readLines output of this data is: readLines("dataset.txt",n=20) [1] " "
2009 May 19
0
File too big for filehash?
Dear R users, I try to use a very large file (~3 Gib) with the filehash package. The length of the dataset is around 4,000,000 obs. I get this message from R while trying to "load" the dataset (named "cc084.csv"): > dumpDF(read.csv("cc084.csv", header=T), dbName="db01") Erreur : impossible d'allouer un vecteur de taille 15.6 Mo (French) Error:
2009 Jan 23
0
Appending objects created using filehash package
Hi, I am working with a very large dataset, and am using the 'filehash' package to manage such a large file. While I have no problem accessing objects that I load into a database, I was hoping there is a better way to append to objects already in the database. The only way I know now to append to an object, basically requires rewriting the entire object. Sample code:
2011 Oct 06
3
Running *slow*
Anyone got any hints on how to make this code more efficient? An early version (which to be fair did more than this one is) ran for 330 hours and produced no output. I have a two column table, Dat, with 12,000,000 rows and I want to produce a lookup table, ltable, in a 1 dimensional matrix with one copy of each of the values in Dat: for (i in 1:nrow(Dat)) { for (j in 1:2) { #If next
2012 Jul 24
4
ERROR : cannot allocate vector of size (in MB & GB)
Hi, Here in R, I need to load a huge file(.csv) , its size is 200MB. [may come more than 1GB sometimes]. When i tried to load into a variable it taking too much of time and after that when i do cbind by groups, getting an error like this " Error: cannot allocate vector of size 82.4 Mb " My requirement is, spilt data from Huge-size-file(.csv) to no. of small csv files. Here i will give
2008 Jul 01
2
"Invalid object" error in boxplot
Hi, I'm trying to make a boxplot with the data at the end of the message, and when I try to execute the command >boxplot(Diatoms) (or for any other field instead of "Diatoms") I get the following error message: Error in oldClass(stats) <- cl : adding class "factor" to an invalid object Any advice would be much appreciated. Thanks a lot, Miriam Date
2003 Jan 24
4
memory problems
Hi I'm computing a bca interval using bca.ci from the boot package. When I try to use this I get an error > library(boot) > boot(logglm.data,boot.fishpower,2500,coef.vec=coeflm.vec)->blm8901 > bca.ci(blm8901,index=29) Error: cannot allocate vector of size 456729 Kb However my machine has 2GB of memory and without R running I only have 112M of memory used. Is there something
2008 Jan 31
4
RSpec-1.1.3
RSpec-1.1.3 has been released. If you''re using RSpec and autotest, you''ll have to upgrade to RSpec-1.1.3 and ZenTest-3.9.0 at the same time. == Version 1.1.3 Maintenance release. * Tightened up exceptions list in autotest/rails_spec. Closes #264. * Applied patch from Ryan Davis for ZenTest-3.9.0 compatibility * Applied patch from Kero to add step_upcoming to story listeners.
2008 Nov 07
6
php-cgi frozen with sbwait when SMP enable
Hello, I have 4 web servers with lighttpd to serve one web site with DNS load sharing. On the 2 SMP-enable web servers, there will be many php-cgi frozen in 'sbwait' state every day. It means the php-cgi stay in 'sbwait' state, and never be back to 'accept' or other state. If I restart them, there will be frozen php-cgi appear some hours later. There is no problem on the
2018 Apr 27
3
Size of produced binaries when compiling llvm & clang sources
Dear llvm developpers, I followed the tutorial to build llvm and clang provided here: https://clang.llvm.org/get_started.html The sources are in sync with subversion repository, and I ended up with more than 30GB of binaries in llvm/bin as shown at the end of this message. I assume I did something wrong, but I did not find any entry in the doc that helps me understand how to reduce the size of
2006 Sep 13
3
FreeBSD 6.1-RELEASE/kqueue high CPU load
Hi to ALL! I have dovecot-1.0r7 installed on FreeBSD 6.1, using kqueue and Maildir (<20 mailboxes, < 0,5 Gb size). Periodically CPU load of imap processes increasing up to 60-80%. Is it normall behavior or not? May be somebody had such kind of problems? -- ? ?????????, ?????? ????? ??? "???????" : ?????????? ??, WEB-?????????? +7 (495) 589 68 81 +7 (926) 575 22 11
2018 Apr 27
0
Size of produced binaries when compiling llvm & clang sources
On Fri, Apr 27, 2018 at 6:21 PM, Manuel Yguel via llvm-dev <llvm-dev at lists.llvm.org> wrote: > Dear llvm developpers, > I followed the tutorial to build llvm and clang provided here: > https://clang.llvm.org/get_started.html > > The sources are in sync with subversion repository, and I ended up with more > than 30GB of binaries in llvm/bin as shown at the end of this