Displaying 20 results from an estimated 140 matches for "30,000".
Did you mean:
40,000
2002 Nov 13
1
building a formula for glm() with 30,000 independent vari ables
...or book?
Thanks in previous,
Dominik
> -----Original Message-----
> From: ripley at stats.ox.ac.uk [mailto:ripley at stats.ox.ac.uk]
> Sent: dimanche, 10. novembre 2002 18:55
> To: Ben Liblit
> Cc: r-help at stat.math.ethz.ch
> Subject: Re: [R] building a formula for glm() with 30,000 independent
> variables
>
>
> Well, the theory of perceptrons says you will find perfect
> discrimination
> with high probability even if there is no structure unless n
> is well in
> excess of 2p. So you do have 100,000 units? If so you have many
> gigabytes of...
2002 Nov 10
7
building a formula for glm() with 30,000 independent variables
I would like to use R to perform a logistic regression with about
30,000 independent variables. That's right, thirty thousand. Most
will be irrelevant: the intent is to use the regression to identify
the few that actually matter.
Among other things, this calls for giving glm() a colossal "y ~ ..."
formula with thirty thousand summed terms on its rig...
2008 May 31
0
Advice for working with Sammon's Projection on image data
Hello all,
I'm working on a project that uses a SOM to generate feature maps from
image data. Each image is a 100x75 pixels and RGBA, so 30,000 elements
per image.
I wanted to observe the structure of the pixel by pixel euclidean
distance (where each image is a point in 30,000 dimensional space) of my
image data.
Sammon's projection seems appropriate for this, though I'm a bit
concerned 400 images with 30,000 dimensions make...
2007 Feb 12
3
processing a large matrix
...formula and looping through every column:
> summary(lm(matrix[,x]~matrix[,y]))$r.squared
If I have 10,000 columns, the loops (10,000 * 10,000) take forever even if
there is no formula inside.
Then, I attempted to vectorize my code:
> cor(matrix)^2
With 10,000 columns, this works great. With 30,000, R tells me it cannot
allocate vector of that length even if the memory limit is set to 4 GBs.
Is there anything else I can do to resolve this issue?
Thanks.
--
View this message in context: http://www.nabble.com/processing-a-large-matrix-tf3216447.html#a8932591
Sent from the R help mailing...
2009 Dec 08
1
Seagate announces enterprise SSD
FYI,
Seagate has announced a new enterprise SSD. The specs appear
to be competitive:
+ 2.5" form factor
+ 5 year warranty
+ power loss protection
+ 0.44% annual failure rate (AFR) (2M hours MTBF, IMHO too low :-)
+ UER 1e-16 (new), 1e-15 (5 years)
+ 30,000/25,000 4 KB read IOPS (peak/aligned zero offset)
+ 30,000/10,500 4 KB write IOPS (peak/aligned zero offset)
http://www.seagate.com/www/en-us/products/servers/pulsar/pulsar/
http://storageeffect.media.seagate.com/2009/12/storage-effect/seagate-pulsar-the-first-enterprise-ready-ssd/
http://www....
2006 Nov 01
3
User quotas. A recurring question
One question that keeps coming up in my discussions about ZFS is the lack of user quotas.
Typically this comes from people who have many tens of thousands (30,000 - 100,000) of users where they feel that having a file system per user will not be manageable. I would agree that today that is the case however I personally don''t see this as an insurmountable problem. However the questions keep being asked:
Could user quotas be introduced on ZFS?
A...
2004 Jul 14
4
duplicate row importing issue
Hello,
I'm simply trying to import a .txt file that has a column of
about 30,000 pts. I found the file, but I'm getting an error:
> m <- read.table(choose.files())
Error in "row.names<-.data.frame"(`*tmp*`, value = row.names) :
duplicate row.names are not allowed
Any help with getting around this?
I really appreciate all the help.
Thanks...
2004 Jul 14
3
(no subject)
...- read.table("/Desktop/work/128_L")
Error in file(file, "r") : unable to open connection
In addition: Warning message:
cannot open file `/Desktop/work/128_L'
do you know why this isn't working? All I have is a bunch of text files,
each with a single column of about 30,000 rows.
thanks
[[alternative HTML version deleted]]
2005 Feb 22
2
High CPU utilization for smbd on AIX
...ning on a windows 2000 server that mapps a
drive to HEATHER. This application will periodically poll the mapped
drive for files (*.sql - small text files with an MSSQL query in them),
read them, execute the query to an MSSQL server and then delete the
file. This directory can have anywhere from 0-30,000 files in it at any
given time. Whenever the process on the windows 2000 server is
accessing the samba drive the smbd process servicing it pretty much eats
up the cpu on the samba server. The process on the win2k server goes
through the files on the samba drive very slowly no matter how many
f...
2009 Aug 30
3
Sapply
Hi,
I need a bit of guidance with the sapply function. I've read the help
page, but am still a bit unsure how to use it.
I have a large data frame with about 100 columns and 30,000 rows. One
of the columns is "group" of which there are about 2,000 distinct "groups".
I want to normalize (sum to 1) one of my variables per-group.
Normally, I would just write a huge "for each" loop, but have read that
is hugely inefficient with R.
The old w...
2007 Jul 23
4
RoR Postion. Full time, Utah
...ons open for experienced ROR programmers. If you feel
like you are a experienced ROR programmer and are looking for a full
time position, please send over you resume. You must be willing to re-
locate. No contract jobs. We have an in house team and have only 2
positions left.
Starting Salary DOE. $30,000-$60,000 with housing, full benefits,
401k, paid vacation, ext. Don''t miss out on this opportunity if you
are able to relocate and are seeking a full time position.
Please send resumes to trevor"dot"milton "@" upillar.com
or........... isebastion"@"yaho...
2007 Jul 24
1
Using lmer with huge amount of data
...ls, a random factor B (subject), and a covariate (age)
with a model like this
MyResult <- lmer(Response ~ FactorA + Age + (1 | subject), MyData, ...)
My input data are like this: For each subject I have a file (a huge
matrix) storing the response values of the subject at many locations
(~30,000 voxels) corresponding to factor A at the 1st level, another
file for factor A at the 2nd level, and a 3rd file for factor A at
the 3rd level. Then I have another file storing the age of those
subjects. The analysis with the linear mixed model above would be
done at each voxel separately...
2011 Sep 19
6
64.31.19.48 attempt to break into my computer
...k pass; user unknown
Sep 19 01:16:45 lin12 dovecot-auth: pam_unix(dovecot:auth): authentication failure; logname= uid=0 euid=0 tty=dovecot ruser= rhost=::ffff:64.31.19.48
Sep 19 01:16:45 lin12 dovecot-auth: pam_succeed_if(dovecot:auth): error retrieving information about user abby
etc. Literally, 30,000 user names attempted.
--
rick baartman
TRIUMF
4004 Wesbrook Mall
Vancouver, BC
V6T2A3
2018 Apr 18
3
Pasar variables de data frame a una sola variable
Hola erreros!
Acudo de nuevo a ustedes porque ando enredado con algo que debe ser simple
pero que no consigo. Lo que ando buscando hacer es pasar todas las
variables de un dataframe a solo una variable (columna) con la info
apilada. Lo que tengo es un dataframe así (con 200 variables y 30,000
filas):
Var1 Var2 Var3
1 5 9
2 6 10
3 7 11
4 8 12
y lo que necesito es convertirlo a:
Var1
1
2
3
4
5
6
7
8
9
10
11
12
Me queda la duda también de cómo se podría hacer a la inversa.
Gracias de antemano por su ayuda!
Diego Iglesias
[[alternative HTML version deleted]]
2013 May 31
1
R 3.0.1 : parallel collection triggers "long memory not supported yet"
...do)
8: FUN(1:3[[3]], ...)
9: sendMaster(try(lapply(X = S, FUN = FUN, ...), silent = TRUE))
Selection: .....................Error in sendMaster(try(lapply(X = S, FUN =
FUN, ...), silent = TRUE)) :
long vectors not supported yet: memory.c:3100
admittedly, my outcome will be a very big list, with 30,000 elements, each
containing data frames with 14 variables and around 200 to 5000
observations (say, 64KB on average). thus, I estimate that the resulting
list is 20GB. the specific code that triggers this is
exposures.list <- mclapply(1:length(crsp.list.by.permno),...
2011 Feb 08
2
Extrcat selected rows from a list
Hi,
I have two lists
1) List1- 30,000 rows and 104 columns
2) List2- a list of 14000 selected rownames from List 1
Now, I want to extract all the 104 columns of List1 matching with the 14000 selected rownames from List2.
Psedocode will be something like this:
match rownames(List2) with rownames(List1)
extract selected match...
2009 Aug 26
3
changing equal values on matrix by same random number
Dear all,
I have about 30,000 matrix (512x512), with values from 1 to N.
Each value on a matrix represent a habitat patch on my
matrix (i.e. my landscape). Non-habitat are stored as ZERO.
No I need to change each 1-to-N values for the same random
number.
Just supose my matrix is:
mymat<-matrix(c(1,1,1,0,0,0,0,0,0,0,0,
0...
2007 Sep 12
3
Document Scanning and Storage
I'd like to start scanning our boxed up documents. I'd say about 30,000
files total.
Mostly to eliminate the boxes of paper we have.
I'd like to scan them, store them, Have some sort of index, and be able to
retrieve them on multiple machines. I think PDF would be the desired format.
I'd like be able to set some permissions as well. (not a deal breaker...)...
2006 Aug 16
6
read.csv issue
...sch_name
1 331-802-7081 School One
2 464-551-7357 School Two
3 388-517-7627 School Three & Four
4 388-517-4394 School Five
It turns out to be very important to read in this character as I have a
program that loops through a data set and Sweave's about 30,000 files.
The variable sch_name gets dropped into the tex file using
\Sexpr{tmp$sch_name}. However, if there is an &, the latex file won't
compile properly. So, what I need is for the data to be read in as
schid sch_name
1 331-802-7081 School One
2 464-551...
2001 Jan 11
4
read data into R with some constraints
Hi,
I have a big data file (over 30,000 records) looks
like this:
100, 20, 46, 70
103, 0, 22, 45
117, -1, 34, 65
120, 15, 0, 25
113, 0, -1, 32
142, -1, -1, 55
.....
I want to read only those records having positive
values in all of the four
columns. That is, I don't want to read record # 3, 5,
and 6 into R. However,
when...