similar to: Understanding lm-based analysis of fractional factorial experiments

Displaying 20 results from an estimated 1000 matches similar to: "Understanding lm-based analysis of fractional factorial experiments"

2002 Dec 12
1
rsync error with novell server
Hi guys! I'm using rsync to update my work and mail from my Unix-workstation to my home directory on a Novell-Server... This is what I get as a message after the rsync is finished rsync error: partial transfer (code 23) at main.c(578) I understand that rsync was trying to chown a directory on the server which Novell doesn't support (or our admin doesn't :}) ... however, I
2010 Feb 28
3
Change the scale on a barplot's y axis
I have grades data. I read them from a csv in letter-grade format. I then converted them to levels levels(grades$grade)=c('A+','A','A-','B+','B','B-','C+','C','C-','D+','D','D-') And then to numbers grades$gp=grades$grade levels(grades$gp)=c(4.3,4.0,3.7, 3.3,3.0,2.7, 2.3,2.0,1.7, 1.3,1.0,0.7)
2010 Mar 06
3
r code to generate interaction columns
Hi, is there a way to take a dataset and extract numeric columns and create interaction columns from it automatically? For e.g. there are 5 columns of data: A,B,C,D,E. CDE are numeric. Can someone provide code to automatically create more columns such as: 1) C*D, C*E, C*D*E, (C+E)/(D+.01 (to avoid divide by zero), (D+E)/(C+.01 (to avoid divide by zero), (C+D)/(E+.01 (to avoid
2009 Mar 23
1
lattice multipanel strip placement - with two factors
Hi, I'm making a multipanel lattice densityplot figure with 2 factors (3 and 20 classes in each factor) with the following statement (the type="percent" is there to prevent plotting the actual points which detract from the figure - is there another way of doing this?): densityplot(~End-Begin | Type * Chromosome, data=Mon, layout=c(5,12), xlab="Element
2006 Aug 30
1
Optimization
Dear R-list, I'm trying to estimate the relative importance of 6 environmental variables in determining clam yield. To estimate clam yield a previous work used the function Yield = (A^a*B^b*C^c...)^1/(a+b+c+...) where A,B,C... are the values of the environmental variables and the weights a,b,c... have not been calibrated on data but taken from literature. Now I'd like to estimate the
2013 Mar 25
1
a contrast question
Dear R People: I have the following in a file: resp factA factB 39.5 low B- 38.6 high B- 27.2 low B+ 24.6 high B+ 43.1 low B- 39.5 high B- 23.2 low B+ 24.2 high B+ 45.2 low B- 33.0 high B- 24.8 low B+ 22.2 high B+ and I construct the data frame: > collard.df <- read.table("collard.txt",header=TRUE) > collard.aov <- aov(resp~factA*factB,data=collard.df) >
2011 Mar 27
6
Asking Favor For the Script of Median Filter
Hello,everybody. My name is Chuan Zun Liang. I come from Malaysia. I am just a beginner for R. Kindly to ask favor about median filter. The problem I facing as below: > x<-matrix(sample(1:30,25),5,5) > x [,1] [,2] [,3] [,4] [,5] [1,] 7 8 30 29 13 [2,] 4 6 12 5 9 [3,] 25 3 22 14 24 [4,] 2 15 26 23 19 [5,] 28 18 10 11 20 This
2006 Aug 17
2
Creating function name from variables..
I have a situation where I have a number of fields and a rating for the field. So, in a DRY way, this works fine: <%= [''abc'', ''def'', ''klm''].each do |e| %> <tr><td><%= @result[e] %></td> <td> <%= @result[e+''rating''] %></td> </tr> <% end %> Now, I
2005 Dec 21
2
Random numbers
Hi All. I have R code whose functionality is being replicated within a C+ program. The outputs are to be compared to validate the conversion somewhat - however (as is always the case) I have stuffed my code with random number calls. Random uniform numbers in C+ are being produced using the (Boost) mersenne-twister generators (mt11213b & mt19937) - which is the default type of generator
2010 Sep 27
7
Regular expressions: offsets of groups
Dear list! > gregexpr("a+(b+)", "abcdaabbc") [[1]] [1] 1 5 attr(,"match.length") [1] 2 4 What I want is the offsets of the matches for the group (b+), i.e. 2 and 7, not the offsets of the complete matches. Is there a way in R to get that? I know about gsubgn and strapply, but they only give me the strings matched by groups not their offsets. I could write
2012 Aug 23
1
All possible models with nls()
Hi all, I am trying to make a script that prints all possible models from a model I've created using nls(). It is a logisitc model which in total includes 13 variables. So its >8000 models I need to create, which I don't want to do manually. I've tried modify scripts made for linear models with no results. I've tried these scripts on a two variable model (c,a1 and a2 is what I
2018 Feb 09
3
[PATCH]Add address overflow check
Hi, I came into a crash when using 32-bit `speexdec` and found that there's an address overflow in function `print_comments()`: static void print_comments(char *comments, int length) { char *c=comments; int len, i, nb_fields; char *end; if (length<8) { fprintf (stderr, "Invalid/corrupted comments\n"); return; } end = c+length;
2006 Jun 15
1
Repost: Estimation when interaction is present: How do I get get the parameters from nlme?
Gday, This is a repost since I only had one direct reply and I remain mystified- This may be stupidity on my part but it may not be so simple. In brief, my problem is I'm not sure how to extract parameter values/effect sizes from a nonlinear regression model with a significant interaction term. My data sets are dose response curves (force and dose) for muscle that also have two
2012 Mar 08
4
Problem with a Matrix
Dear All, I have a problem in making a matrix from a data. I did a summary data from a big data frame using (ddply), the resulting table is like that: Site.No Plant.Name Cover.Percentage 1 XXX 15 1 YYY 20 2 XXX 18 2 YYY 21 3 XXX 14 3
2011 Jun 03
2
Y error bars, dates on the X-axis
Dear forum, Does anyone know any functions for plotting vertical error bars, if the x-axis variable is a date? I can produce a pretty plot with dates on the x-axis, if I don't want to include error bars. I can also produce a nice plot with error bars, if I just use Julian days for the x-axis, but I'd much rather have the actual dates. I have tried writing my own function, plotCI
2003 Jul 25
0
problem with full partition
Hi, i have problem with my redhat 7.3 I have extra HDD which is full AND it doesn't show me my files any more!!! I add some explanations in the end. how can i get back my files? thanks, riho ------------------ [root@bla minudata]# df Filesystem 1k-blocks Used Available Use% Mounted on /dev/hda1 1031800 926496 52892 95% / /dev/hda3
2007 Jun 19
2
Traktor-Icecast
I have Traktor DJ studio 3.2 how do i get Icecast to work with it because it has totally confused me? Ryan Hussey (r.hussey@hotmail.co.uk) _________________________________________________________________ Feel like a local wherever you go with BackOfMyHand.com http://www.backofmyhand.com -------------- next part -------------- An HTML attachment was scrubbed... URL:
2013 Feb 06
2
The interpretation of lm(y~x)?
Hi, I am reading the book "Mixed Effects Models in S and S-Plus" and come across an example with the Rail data. I tried to use lm(travel~Rail,data=Rail) and got the following result: Call: lm(formula = travel ~ Rail, data = Rail) Residuals: Min 1Q Median 3Q Max -6.6667 -1.0000 0.1667 1.0000 6.3333 Coefficients: Estimate Std. Error t value Pr(>|t|)
2007 Apr 18
7
[RFC, PATCH 5/24] i386 Vmi code patching
The VMI ROM detection and code patching mechanism is illustrated in setup.c. There ROM is a binary block published by the hypervisor, and and there are certainly implications of this. ROMs certainly have a history of being proprietary, very differently licensed pieces of software, and mostly under non-free licenses. Before jumping to the conclusion that this is a bad thing, let us consider more
2007 Apr 18
7
[RFC, PATCH 5/24] i386 Vmi code patching
The VMI ROM detection and code patching mechanism is illustrated in setup.c. There ROM is a binary block published by the hypervisor, and and there are certainly implications of this. ROMs certainly have a history of being proprietary, very differently licensed pieces of software, and mostly under non-free licenses. Before jumping to the conclusion that this is a bad thing, let us consider more