similar to: Can I build an array of regrssion model?

Displaying 20 results from an estimated 4000 matches similar to: "Can I build an array of regrssion model?"

2002 Dec 25
2
Is there string class in R?
Hi, I want to use R handle hundreds data file. So I want use some string class to handle the names of data file. My idea procedure like: for i in 1:20 { filename = "***" + i +".PRN" scan(filename) .... } Do anyone know how to do this? Thanks,
2003 Sep 28
1
geodata conversion
Hey, I have a data set of 1.20MB. I used read.table( ) to read in the data and then tried to convert it to geodata type. But it failed and I got "Error in vector("double", length) : cannot allocate vector of length 1351974000" message. All the commands I used were: rawdata<-read.table(file=" ") datag<-as.geodata(rawdata) BTW, what is the biggest origin
2011 May 17
1
simprof test using jaccard distance
Dear All, I would like to use the simprof function (clustsig package) but the available distances do not include Jaccard distance, which is the most appropriate for pres/abs community data. Here is the core of the function: > simprof function (data, num.expected = 1000, num.simulated = 999, method.cluster = "average", method.distance = "euclidean", method.transform =
2012 Dec 10
1
Can somebody suggest how to achieve following data manipulation?
Dear all, Let say I have following data: RawData <- matrix(1:101, nr = 1); colnames(RawData) <- c("ASD", as.character(as.yearmon(seq(as.Date("2012-03-01"), length.out = 100, by = "1 month")))); rownames(RawData) <- "XYZ" CutOffDate <- as.Date("2012-09-01") NewDateSeries <- as.character(as.yearmon(seq(CutOffDate, to =
2008 Mar 23
1
mapply
In an earlier post, a person wanted to divide each of the rows of rawdata by the row vector sens so he did below but didn't like it and asked if there was a better solution. rawdata <- data.frame(rbind(c(1,2,2), c(4,5,6))) sens <- c(2,4,6) temp <- t(rawdata)/sens temp <- t(temp) print(temp) Gabor sent three other solutions and I understood 2 of them but not the
2005 Aug 04
2
Avoiding for loop
I understand that in R, for loops are not used as often as other languages, and am trying to learn how to avoid them. I am wondering if there is a more efficient way to write a certain piece of code, which right now I can only envision as a for loop. I have a data file that basically looks like: 1,55 1,23 2,12 ... that defines a matrix. Each row of the data file corresponds to a row of the
2006 Jul 12
4
Keep value lables with data frame manipulation
Dear R, I import data from spss into a R data.frame. On this rawdata I do some data processing (selection of observations, normalization, recoding of variables etc..). The result is stored in a new data.frame, however, in this new data.frame the value labels are lost. Example of what I do in code: # read raw data from spss rawdata <- read.spss("./data/T50937.SAV",
2009 Jul 14
2
How to provide list as an argument for the data.frame()
Hi R -users, i've a table as describe below. I'm reading the numeric value presented in this table to populate a list. #table #============ #X    A    B    C #x1    2    3    4 #x2    5    7    10 #x4    2    3    5 #============ rawData <- read.table("raw_data.txt",header=T, sep="\t") myList=list() counter=0 for (i in c(1:length(rawData$X))) {     print (i)    
2012 Jan 20
2
rbind()
Hello there, Much thanks in advance for any help. I have a few questions: 1) Why do I keep getting the following error: File1 <- read.csv("../RawData/File1.csv",as.is=TRUE,row.names=1) Error in file(file, "rt") : cannot open the connection In addition: Warning message: In file(file, "rt") : cannot open file '../RawData/File1.csv': No such file or
2011 Sep 08
2
pie chart
Hi All, I have txt file like : $ cat data.txt US 10 UK 12 Ind 4 Germany 14 France 8 > rawdata <- read.table(file='data.txt',sep='\t' , header=FALSE) > rawdata V1 V2 1 US 10 2 UK 12 3 Ind 4 4 Germany 14 5 France 8 I want to draw pie chart for the above data. How to split rawdata into : con <-
2009 Aug 03
2
Scale set of 0 values returns NAN??
Hi, More questions in my ongoing quest to convert from RapidMiner to R. One thing has become VERY CLEAR: None of the issues I'm asking about here are addressed in RapidMiner. How it handles misisng values, scaling, etc. is hidden within the "black box". Using R is forcing me to take a much deeper look at my data and how my experiments are constructed. (That's a very
2009 Sep 28
1
help with lda function
I am having a problem understanding the lda package. I have a dataset here: [,1] [,2] [,3] [1,] 2.95 6.63 0 [2,] 2.53 7.79 0 [3,] 3.57 5.65 0 [4,] 3.16 5.47 0 [5,] 2.58 4.46 1 [6,] 2.16 6.22 1 [7,] 3.27 3.52 1 If I do the following; "names(d)<-c("y","x1","x2") d$x1 = d$x1 * 100 d$x2 = d$x2 * 100 g<-lda( y ~ x1 + x2, data=d) v2
2012 Jun 28
1
Merging listed dataset into one
Hello, I'm wondering how I can merge two featuresets into one. My dataset is two sets of microarray data and it looks like followings: > rawData $v1 TilingFeatureSet (storageMode: lockedEnvironment) assayData: 2197815 features, 59 samples element names: channel1, channel2 protocolData rowNames: LT290677RU_D1_2011-02-16 LT286300LU_D1_2010-07-24 ... LT003990RU_D1_2010-11-04 (59
2009 Jun 16
1
adressing dataframes
Hi everyone, I experience some problems with adressing of data.frames when I retrieve some information for geographical position (ypos, xpos) ot of a MySQL Database and want to perform some simple statistics. The problem is adressing the dataframes with a construct like rawdata[c(type)] vs. rawdata$TEMPMIN to retrieve the numerical information and not a string (I want to store the numerical
2003 Dec 10
2
OT: BibTex year-only citation in text?
Sorry for the off-topic question, but I know there are some talented LaTeX users out there. Which bibliography style gives only the year in text citations (e.g "for further details, see Anderson (1992)" )? Thanks Jason -- Indigo Industrial Controls Ltd. http://www.indigoindustrial.co.nz 64-21-343-545 jasont at indigoindustrial.co.nz
2003 Oct 21
5
run R under linux
Dear all, Our department uses the linux system and we are not allowed to submit job directly. We must make a batch to submit through "qmon". so, I make a foo.sh file, which only contains one line: nohup R --vanilla < foo.txt > foo.results foo is all my codes. It is a simulation of 200 times. I set the seed at the beginning. It is to estimate the success probability, which is
2008 Mar 22
2
More elegant multiplication or division of a data frame with a vector
Hello, I am importing some raw voltage multichannel measurements into an R data frame. I need to scale each column with the respective sensitivity for that channel. I figured how to do it, but I am curious if there isn't a more elegant way. Now I start with something like this: rawdata <- data.frame(rbind(c(1,2,3), c(4,5,6))) sens <- c(2,4,6) and I do this: data <-
2009 Sep 20
1
Return a list from a .Call but segfaults
Hello, I call a function via .Call passing to it a raw vector(D) and an integer(I) The vector is a series K1,KData1, V1,VData1, K2, KData2, ... where the integer K1 is the length of Data1 and similarly for Ki (wrt Datai)(similarly for V*) There 2*I such pairs( (Ki,KDatai), (Vi,VDatai)) The numbers Ki(and Vi) are written in network order. I am returning a list of I elements each element a
2003 Nov 01
2
Question about the high dimensional density estimation
Hi, I found that the R package "KernSmooth" can deal with only 1D and 2D data. But now I have a collection of 4-dimensional data (x1,x2,x3,x4) and would like to estimate the "mode" of the underlying density. What can I do about it ? Thanks a lot. -- Ying-Chao Hung Assistant Professor Graduate Institute of Statistics National Central University Chung-Li, Taiwan TEL:
2009 Sep 29
1
help with lda function from MASS package
Thanks David, Yes, I am talking about the MASS package.Thank you for pointing out that these scale the same. My question is, how do I get from the V1 data: V1 1 164.4283 2 166.2492 3 170.5232 4 156.5622 5 127.7540 6 136.7704 7 136.3436 to the other set of data: + 1 -2.3769280 + 2 -2.7049437 + 3 -3.4748309 + 4 -0.9599825 + 5 4.2293774 + 6 2.6052193 + 7 2.6820884 On Mon, Sep 28, 2009