Displaying 20 results from an estimated 20000 matches similar to: "Bug in predict(newdata=x) with poly() (PR#1258)"
2006 Sep 28
2
safe prediction from lm
I am fitting a regression model with a bs term and then making predictions
based on the model. According to some info on the internet at
http://www.stat.auckland.ac.nz/~yee/smartpred/DummiesGuide.txt
there are some problems with using predict.lm when you have a model with
terms such as bs,ns,or poly. However when I used one of the examples they
said would illustrate the problems I get virtually
2003 Oct 27
2
problem using do.call and substitute for predict.glm using poly()
Hi
I am having a particular problem with some glm models I am running. I
have been adapting code from Bill Venables 'Programmers niche' in RNews
Vol 2/2 to fit ca. 1000 glm models to a combination of species 0/1 data
(as Y) and related physicochemical data (X), to automate the process of
fitting this many models. I have successfully managed to fit all the
models and have stored the
2007 May 09
1
predict.tree
I have a classification tree model similar to the following (slightly
simplified here):
> treemod<-tree(y~x)
where y is a factor and x is a matrix of numeric predictors. They have
dimensions:
> length(y)
[1] 1163
> dim(x)
[1] 1163 75
I?ve evaluated the tree model and am happy with the fit. I also have a
matrix of cases that I want to use the tree model to classify. Call it
2013 Nov 14
1
issues with calling predict.coxph.penal (survival) inside a function
Thanks for the reproducable example. I can confirm that it fails on my machine using
survival 2-37.5, the next soon-to-be-released version,
The issue is with NextMethod, and my assumption that the called routine inherited
everything from the parent, including the environment chain. A simple test this AM showed
me that the assumption is false. It might have been true for Splus. Working this
2006 May 27
1
Recommended package nlme: bug in predict.lme when an independent variable is a polynomial (PR#8905)
Full_Name: Renaud Lancelot
Version: Version 2.3.0 (2006-04-24)
OS: MS Windows XP Pro SP2
Submission from: (NULL) (82.239.219.108)
I think there is a bug in predict.lme, when a polynomial generated by poly() is
used as an explanatory variable, and a new data.frame is used for predictions. I
guess this is related to * not * using, for predictions, the coefs used in
constructing the orthogonal
2013 Apr 01
2
example to demonstrate benefits of poly in regression?
Here's my little discussion example for a quadratic regression:
http://pj.freefaculty.org/R/WorkingExamples/regression-quadratic-1.R
Students press me to know the benefits of poly() over the more obvious
regression formulas.
I think I understand the theory on why poly() should be more numerically
stable, but I'm having trouble writing down an example that proves the
benefit of this.
I
2010 Jan 18
2
Predict polynomial problem
I have a function that fits polynomial models for the orders in n:
lmn <- function(d,n){
models=list()
for(i in n){
models[[i]]=lm(y~poly(x,i),data=d)
}
return(models)
}
My data is:
> d=data.frame(x=1:10,y=runif(10))
So first just do it for a cubic:
> mmn = lmn(d,3)
> predict(mmn[[3]])
1 2 3 4 5 6 7 8
2012 Dec 03
2
Different results from random.Forest with test option and using predict function
Hello R Gurus,
I am perplexed by the different results I obtained when I ran code like
this:
set.seed(100)
test1<-randomForest(BinaryY~., data=Xvars, trees=51, mtry=5, seed=200)
predict(test1, newdata=cbind(NewBinaryY, NewXs), type="response")
and this code:
set.seed(100)
test2<-randomForest(BinaryY~., data=Xvars, trees=51, mtry=5, seed=200,
xtest=NewXs, ytest=NewBinarY)
The
2003 May 08
1
predict function (PR#2958)
Full_Name: Murray H Smith
Version: 1.6.1
OS: Windows
Submission from: (NULL) (202.36.29.1)
This is report is more of a matter of completeness rather than an outright bug.
The predict function does not handle the prediction from the constant model
appropriately. It also differs from Splus in this respect.
The length of the vector (or first dimension of the matrix, if type = "terms" is
2015 Apr 30
2
predict nlme
Estimado Oliver Nuñez
Envío un ejemplo reproducible.
Javier Marcuzzi
# de donde tomo datos, y tiene el modelo (en el pdf)
library(MCMCglmm)
# librería con las funciónes que voy a usar
library(nlme)
datos0<-ChickWeight
# creo algunos datos que agrego a los origonales
Factor<-as.numeric(datos0$Chick)
Factor[Factor > 0 & Factor <= 10] <- 'A'
Factor[Factor > 10
2012 Mar 14
0
using predict() with poly(x, raw=TRUE)
Dear r-devel list members,
I've recently encountered the following problem using predict() with a model
that has raw-polynomial terms. (Actually, I encountered the problem using
model.frame(), but the source of the error is the same.) The problem is
technical and concerns the design of poly(), which is why I'm sending this
message to r-devel rather than r-help.
To illustrate:
2009 Jul 13
0
problem predict/poly
Dear R experts,
I am observing undesired behavior of predict(fit, newdata), in case when fit object is produced by lm() involving a poly(). Here is how to reproduce:
x <- c(1:10)
y <- sin(c(1:10))
fit <- lm(formula=y~poly(x, 5, raw=TRUE))
predict(fit, newdata=data.frame(x=c(1:10))) ## this works
predict(fit, newdata=data.frame(x=c(1:1))) ## this is broken, error below
Error in poly(x,
2010 Aug 13
2
Unable to retrieve residual sum of squares from nls output
Colleagues,
I am using "nls" successfully (2.11.1, OS X) but I am having difficulties retrieving part of the output - residual sum of squares. I have assigned the output to FIT:
> > FIT
> Nonlinear regression model
> model: NEWY ~ PMESOR + PAMPLITUDE * cos(2 * pi * (NEWX - POFFSET)/PERIOD)
> data: parent.frame()
> PMESOR PAMPLITUDE POFFSET
>
2013 May 17
2
zigzag confidence interval in a plot
Dear All,
When I plot the values and linear regression line for one data set, it is fine. But for another one I see zigzags, when I plot the confidence interval
>cd
Depth CHAOsep12RNA
9,94 804
25,06 1476,833333
40,04 1540,561404
50,11 1575,166667
52,46 349,222222
54,92 1941,5
57,29 1053,507042
60,11 1535,1
70,04 2244,963303
79,97 1954,507042
100,31 2679,140625
>
2013 Mar 01
1
predict.loess() segfaults for large n?
Hi,
I am segfaulting when using predict.loess() (checked with r62092).
I've traced the source with the help of valgrind (output pasted
below) and it appears that this is due to int overflow when
allocating an int work array in loess_workspace():
liv = 50 + ((int)pow((double)2, (double)D) + 4) * nvmax + 2 * N;
where liv is an (global) int. For D=1 (one x variable), this
overflows at
2005 Feb 15
3
using poly in a linear regression in the presence of NA f ails (despite subsetting them out)
This smells like a bug to me. The error is triggered by the line:
variables <- eval(predvars, data, env)
inside model.frame.default(). At that point, na.action has not been
applied, so poly() ended being called on data that still contains missing
values. The qr() that issued the error is for generating the orthogonal
basis when evaluating poly(), not for fitting the linear model itself.
2005 Feb 15
3
using poly in a linear regression in the presence of NA f ails (despite subsetting them out)
This smells like a bug to me. The error is triggered by the line:
variables <- eval(predvars, data, env)
inside model.frame.default(). At that point, na.action has not been
applied, so poly() ended being called on data that still contains missing
values. The qr() that issued the error is for generating the orthogonal
basis when evaluating poly(), not for fitting the linear model itself.
2013 Feb 24
1
Use of the newdata parameter in the predict.coxph function
Hello,
I've executed the following predict.coxph function to enable prediction for
new variable values (error is included).
*predict(cox_out,newdata=data.frame(Meter3.Value=100.001,
Meter4.Value=200.001,Meter5.Value=300.001,Meter10.Value=
400.001,type="expected"))
Error in model.frame.default(data = data.frame(Meter3.Value = 100.001, :
variable lengths differ (found for
2008 Feb 13
1
use of poly()
Hi,
I am curious about how to interpret the results of a polynomial regression--
using poly(raw=TRUE) vs. poly(raw=FALSE).
set.seed(123456)
x <- rnorm(100)
y <- jitter(1*x + 2*x^2 + 3*x^3 , 250)
plot(y ~ x)
l.poly <- lm(y ~ poly(x, 3))
l.poly.raw <- lm(y ~ poly(x, 3, raw=TRUE))
s <- seq(-3, 3, by=0.1)
lines(s, predict(l.poly, data.frame(x=s)), col=1)
lines(s,
2011 Apr 18
2
Predicting with a principal component regression model: "non-conformable arguments" error
Hello all,
I have generated a principal components regression model using the pcr()
function from the PLS package (R version 2.12.0). I am getting a
"non-conformable arguments" error when I try to use the predict() function
on new data, but only when I try to read in the new data from a separate
file.
More specifically, when my data looks like this
#########training data