search for: 30,000

Displaying 20 results from an estimated 140 matches for "30,000".

Did you mean: 40,000
2002 Nov 13
1
building a formula for glm() with 30,000 independent vari ables
...or book? Thanks in previous, Dominik > -----Original Message----- > From: ripley at stats.ox.ac.uk [mailto:ripley at stats.ox.ac.uk] > Sent: dimanche, 10. novembre 2002 18:55 > To: Ben Liblit > Cc: r-help at stat.math.ethz.ch > Subject: Re: [R] building a formula for glm() with 30,000 independent > variables > > > Well, the theory of perceptrons says you will find perfect > discrimination > with high probability even if there is no structure unless n > is well in > excess of 2p. So you do have 100,000 units? If so you have many > gigabytes of...
2002 Nov 10
7
building a formula for glm() with 30,000 independent variables
I would like to use R to perform a logistic regression with about 30,000 independent variables. That's right, thirty thousand. Most will be irrelevant: the intent is to use the regression to identify the few that actually matter. Among other things, this calls for giving glm() a colossal "y ~ ..." formula with thirty thousand summed terms on its rig...
2008 May 31
0
Advice for working with Sammon's Projection on image data
Hello all, I'm working on a project that uses a SOM to generate feature maps from image data. Each image is a 100x75 pixels and RGBA, so 30,000 elements per image. I wanted to observe the structure of the pixel by pixel euclidean distance (where each image is a point in 30,000 dimensional space) of my image data. Sammon's projection seems appropriate for this, though I'm a bit concerned 400 images with 30,000 dimensions make...
2007 Feb 12
3
processing a large matrix
...formula and looping through every column: > summary(lm(matrix[,x]~matrix[,y]))$r.squared If I have 10,000 columns, the loops (10,000 * 10,000) take forever even if there is no formula inside. Then, I attempted to vectorize my code: > cor(matrix)^2 With 10,000 columns, this works great. With 30,000, R tells me it cannot allocate vector of that length even if the memory limit is set to 4 GBs. Is there anything else I can do to resolve this issue? Thanks. -- View this message in context: http://www.nabble.com/processing-a-large-matrix-tf3216447.html#a8932591 Sent from the R help mailing...
2009 Dec 08
1
Seagate announces enterprise SSD
FYI, Seagate has announced a new enterprise SSD. The specs appear to be competitive: + 2.5" form factor + 5 year warranty + power loss protection + 0.44% annual failure rate (AFR) (2M hours MTBF, IMHO too low :-) + UER 1e-16 (new), 1e-15 (5 years) + 30,000/25,000 4 KB read IOPS (peak/aligned zero offset) + 30,000/10,500 4 KB write IOPS (peak/aligned zero offset) http://www.seagate.com/www/en-us/products/servers/pulsar/pulsar/ http://storageeffect.media.seagate.com/2009/12/storage-effect/seagate-pulsar-the-first-enterprise-ready-ssd/ http://www....
2006 Nov 01
3
User quotas. A recurring question
One question that keeps coming up in my discussions about ZFS is the lack of user quotas. Typically this comes from people who have many tens of thousands (30,000 - 100,000) of users where they feel that having a file system per user will not be manageable. I would agree that today that is the case however I personally don''t see this as an insurmountable problem. However the questions keep being asked: Could user quotas be introduced on ZFS? A...
2004 Jul 14
4
duplicate row importing issue
Hello, I'm simply trying to import a .txt file that has a column of about 30,000 pts. I found the file, but I'm getting an error: > m <- read.table(choose.files()) Error in "row.names<-.data.frame"(`*tmp*`, value = row.names) : duplicate row.names are not allowed Any help with getting around this? I really appreciate all the help. Thanks...
2004 Jul 14
3
(no subject)
...- read.table("/Desktop/work/128_L") Error in file(file, "r") : unable to open connection In addition: Warning message: cannot open file `/Desktop/work/128_L' do you know why this isn't working? All I have is a bunch of text files, each with a single column of about 30,000 rows. thanks [[alternative HTML version deleted]]
2005 Feb 22
2
High CPU utilization for smbd on AIX
...ning on a windows 2000 server that mapps a drive to HEATHER. This application will periodically poll the mapped drive for files (*.sql - small text files with an MSSQL query in them), read them, execute the query to an MSSQL server and then delete the file. This directory can have anywhere from 0-30,000 files in it at any given time. Whenever the process on the windows 2000 server is accessing the samba drive the smbd process servicing it pretty much eats up the cpu on the samba server. The process on the win2k server goes through the files on the samba drive very slowly no matter how many f...
2009 Aug 30
3
Sapply
Hi, I need a bit of guidance with the sapply function. I've read the help page, but am still a bit unsure how to use it. I have a large data frame with about 100 columns and 30,000 rows. One of the columns is "group" of which there are about 2,000 distinct "groups". I want to normalize (sum to 1) one of my variables per-group. Normally, I would just write a huge "for each" loop, but have read that is hugely inefficient with R. The old w...
2007 Jul 23
4
RoR Postion. Full time, Utah
...ons open for experienced ROR programmers. If you feel like you are a experienced ROR programmer and are looking for a full time position, please send over you resume. You must be willing to re- locate. No contract jobs. We have an in house team and have only 2 positions left. Starting Salary DOE. $30,000-$60,000 with housing, full benefits, 401k, paid vacation, ext. Don''t miss out on this opportunity if you are able to relocate and are seeking a full time position. Please send resumes to trevor"dot"milton "@" upillar.com or........... isebastion"@"yaho...
2007 Jul 24
1
Using lmer with huge amount of data
...ls, a random factor B (subject), and a covariate (age) with a model like this MyResult <- lmer(Response ~ FactorA + Age + (1 | subject), MyData, ...) My input data are like this: For each subject I have a file (a huge matrix) storing the response values of the subject at many locations (~30,000 voxels) corresponding to factor A at the 1st level, another file for factor A at the 2nd level, and a 3rd file for factor A at the 3rd level. Then I have another file storing the age of those subjects. The analysis with the linear mixed model above would be done at each voxel separately...
2011 Sep 19
6
64.31.19.48 attempt to break into my computer
...k pass; user unknown Sep 19 01:16:45 lin12 dovecot-auth: pam_unix(dovecot:auth): authentication failure; logname= uid=0 euid=0 tty=dovecot ruser= rhost=::ffff:64.31.19.48 Sep 19 01:16:45 lin12 dovecot-auth: pam_succeed_if(dovecot:auth): error retrieving information about user abby etc. Literally, 30,000 user names attempted. -- rick baartman TRIUMF 4004 Wesbrook Mall Vancouver, BC V6T2A3
2018 Apr 18
3
Pasar variables de data frame a una sola variable
Hola erreros! Acudo de nuevo a ustedes porque ando enredado con algo que debe ser simple pero que no consigo. Lo que ando buscando hacer es pasar todas las variables de un dataframe a solo una variable (columna) con la info apilada. Lo que tengo es un dataframe así (con 200 variables y 30,000 filas): Var1 Var2 Var3 1 5 9 2 6 10 3 7 11 4 8 12 y lo que necesito es convertirlo a: Var1 1 2 3 4 5 6 7 8 9 10 11 12 Me queda la duda también de cómo se podría hacer a la inversa. Gracias de antemano por su ayuda! Diego Iglesias [[alternative HTML version deleted]]
2013 May 31
1
R 3.0.1 : parallel collection triggers "long memory not supported yet"
...do) 8: FUN(1:3[[3]], ...) 9: sendMaster(try(lapply(X = S, FUN = FUN, ...), silent = TRUE)) Selection: .....................Error in sendMaster(try(lapply(X = S, FUN = FUN, ...), silent = TRUE)) : long vectors not supported yet: memory.c:3100 admittedly, my outcome will be a very big list, with 30,000 elements, each containing data frames with 14 variables and around 200 to 5000 observations (say, 64KB on average). thus, I estimate that the resulting list is 20GB. the specific code that triggers this is exposures.list <- mclapply(1:length(crsp.list.by.permno),...
2011 Feb 08
2
Extrcat selected rows from a list
Hi, I have two lists 1) List1- 30,000 rows and 104 columns 2) List2- a list of 14000 selected rownames  from List 1 Now, I want to extract all the 104 columns of List1  matching with the 14000 selected rownames from List2.  Psedocode will be something like this: match rownames(List2) with rownames(List1) extract selected match...
2009 Aug 26
3
changing equal values on matrix by same random number
Dear all, I have about 30,000 matrix (512x512), with values from 1 to N. Each value on a matrix represent a habitat patch on my matrix (i.e. my landscape). Non-habitat are stored as ZERO. No I need to change each 1-to-N values for the same random number. Just supose my matrix is: mymat<-matrix(c(1,1,1,0,0,0,0,0,0,0,0, 0...
2007 Sep 12
3
Document Scanning and Storage
I'd like to start scanning our boxed up documents. I'd say about 30,000 files total. Mostly to eliminate the boxes of paper we have. I'd like to scan them, store them, Have some sort of index, and be able to retrieve them on multiple machines. I think PDF would be the desired format. I'd like be able to set some permissions as well. (not a deal breaker...)...
2006 Aug 16
6
read.csv issue
...sch_name 1 331-802-7081 School One 2 464-551-7357 School Two 3 388-517-7627 School Three & Four 4 388-517-4394 School Five It turns out to be very important to read in this character as I have a program that loops through a data set and Sweave's about 30,000 files. The variable sch_name gets dropped into the tex file using \Sexpr{tmp$sch_name}. However, if there is an &, the latex file won't compile properly. So, what I need is for the data to be read in as schid sch_name 1 331-802-7081 School One 2 464-551...
2001 Jan 11
4
read data into R with some constraints
Hi, I have a big data file (over 30,000 records) looks like this: 100, 20, 46, 70 103, 0, 22, 45 117, -1, 34, 65 120, 15, 0, 25 113, 0, -1, 32 142, -1, -1, 55 ..... I want to read only those records having positive values in all of the four columns. That is, I don't want to read record # 3, 5, and 6 into R. However, when...