Displaying 20 results from an estimated 10000 matches similar to: "survreg with frailty"
2003 May 07
0
Re: frailty models in survreg() -- survival package (PR#2934)
On Tue, 6 May 2003, Jerome Asselin wrote:
>
> I am confused on how the log-likelihood is calculated in a parametric
> survival problem with frailty. I see a contradiction in the frailty() help
> file vs. the source code of frailty.gamma(), frailty.gaussian() and
> frailty.t().
>
> The function frailty.gaussian() appears to calculate the penalty as the
> negative
2003 May 07
0
Re: frailty models in survreg() -- survival package (PR#2934)
SEE ALSO ORIGINAL POSTING IN PR#2933
On May 6, 2003 03:58 pm, Thomas Lumley wrote:
>
> Looking at a wider context in the code
>
> pfun <- function(coef, theta, ndeath) {
> if (theta == 0)
> list(recenter = 0, penalty = 0, flag = TRUE)
> else {
> recenter <- log(mean(exp(coef)))
> coef <- coef - recenter
2003 May 07
0
frailty models in survreg() -- survival package (PR#2933)
I am confused on how the log-likelihood is calculated in a parametric
survival problem with frailty. I see a contradiction in the frailty() help
file vs. the source code of frailty.gamma(), frailty.gaussian() and
frailty.t().
The function frailty.gaussian() appears to calculate the penalty as the
negative log-density of independent Gaussian variables, as one would
expect:
>
2007 Jan 22
0
[UNCLASSIFIED] predict.survreg() with frailty term and newdata
Dear All,
I am attempting to make predictions based on a survreg() model with some censoring and a frailty term, as below: predict works fine on the original data, but not if I specify newdata.
# a model with groups as fixed effect
model1 <- survreg(Surv(y,cens)~ x1 + x2 + groups,
dist = "gaussian")
# and with groups as a random effect
fr <- frailty(groups,
2011 Jan 28
1
survreg 3-way interaction
> I was wondering why survreg (in survival package) can not handle
> three-way interactions. I have an AFT .....
You have given us no data to diagnose your problem. What do you mean
by "cannot handle" -- does the package print a message "no 3 way
interactions", gives wrong answers, your laptop catches on fire when you
run it, ....?
Also, make sure you read
2013 Nov 04
0
Fwd: Re: How to obtain nonparametric baseline hazard estimates in the gamma frailty model?
-------- Original Message --------
Subject: Re: How to obtain nonparametric baseline hazard estimates in the gamma frailty model?
Date: Mon, 04 Nov 2013 17:27:04 -0600
From: Terry Therneau <therneau.terry at mayo.edu>
To: Y <yuhanusa at gmail.com>
The cumulative hazard is just -log(sfit$surv).
The hazard is essentially a density estimate, and that is much harder. You'll notice
2011 Apr 08
1
Variance of random effects: survreg()
I have the following questions about the variance of the random effects in the survreg() function in the survival package:
1) How can I extract the variance of the random effects after fitting a model?
For example:
set.seed(1007)
x <- runif(100)
m <- rnorm(10, mean = 1, sd =2)
mu <- rep(m, rep(10,10))
test1 <- data.frame(Time = qsurvreg(x, mean = mu, scale= 0.5, distribution =
2008 Apr 17
1
survreg() with frailty
Dear R-users,
I have noticed small discrepencies in the reported estimate of the
variance of the frailty by the print method for survreg() and the
'theta' component included in the object fit:
# Examples in R-2.6.2 for Windows
library(survival) # version 2.34-1 (2008-03-31)
# discrepancy
fit1 <- survreg(Surv(time, status) ~ rx + frailty(litter), rats)
fit1
fit1$history[[1]]$theta
2008 Apr 25
3
Use of survreg.distributions
Dear R-user:
I am using survreg(Surv()) for fitting a Tobit model of left-censored longitudinal data. For logarithmic transformation of y data, I am trying use survreg.distributions in the following way:
tfit=survreg(Surv(y, y>=-5, type="left")~x + cluster(id), dist="gaussian", data=y.data, scale=0, weights=w)
my.gaussian<-survreg.distributions$gaussian
2007 Jul 12
1
p-value from survreg
The question was how to get the p-value from the fit below, as an S object
sr<-survreg(s~groups, dist="gaussian")
Coefficients:
(Intercept) groups
-0.02138485 0.03868351
Scale= 0.01789372
Loglik(model)= 31.1 Loglik(intercept only)= 25.4
Chisq= 11.39 on 1 degrees of freedom, p= 0.00074
n= 16
----
In general, good places to start are
> names(sr)
>
2006 Feb 28
1
ex-Gaussian survival distribution
Dear R-Helpers,
I am hoping to perform survival analyses using the "ex-Gaussian"
distribution.
I understand that the ex-Gaussian is a convolution of exponential and
Gaussian
distributions for survival data.
I checked the "survreg.distributions" help and saw that it is possible to
mix
pre-defined distributions. Am I correct to think that the following code
makes
the
2008 Dec 23
6
Interval censored Data in survreg() with zero values!
Hello,
I have interval censored data, censored between (0, 100). I used the
tobit function in the AER package which in turn backs on survreg.
Actually I'm struggling with the distribution. Data is asymmetrically
distributed, so first choice would be a Weibull distribution.
Unfortunately the Weibull doesn't allow for zero values in time data,
as it requires x > 0. So I tried the
2011 Jan 10
4
Meaning of pterms in survreg object?
I am trying to model survival data with a Weibull distribution
using survreg. Units are clustered two apiece, sometimes receiving
the same treatment and sometimes opposing treatment.
2007 Apr 20
1
Approaches of Frailty estimation: coxme vs coxph(...frailty(id, dist='gauss'))
Dear List,
In documents (Therneau, 2003 : On mixed-effect cox
models, ...), as far as I came to know, coxme penalize
the partial likelihood (Ripatti, Palmgren, 2000) where
as frailtyPenal (in frailtypack package) uses the
penalized the full likelihood approach (Rondeau et al,
2003).
How, then, coxme and coxph(...frailty(id,
dist='gauss')) differs? Just the coding algorithm, or
in
2009 Jan 02
0
[Fwd: Re: Interval censored Data in survreg() with zero values!]
-------------- next part --------------
An embedded message was scrubbed...
From: Terry Therneau <therneau at mayo.edu>
Subject: Re: Interval censored Data in survreg() with zero values!
Date: Tue, 30 Dec 2008 16:46:37 -0600 (CST)
Size: 4268
URL: <https://stat.ethz.ch/pipermail/r-help/attachments/20090102/abe75d02/attachment-0002.eml>
2005 May 03
2
comparing lm(), survreg( ... , dist="gaussian") and survreg( ... , dist="lognormal")
Dear R-Helpers:
I have tried everything I can think of and hope not to appear too foolish
when my error is pointed out to me.
I have some real data (18 points) that look linear on a log-log plot so I
used them for a comparison of lm() and survreg. There are no suspensions.
survreg.df <- data.frame(Cycles=c(2009000, 577000, 145000, 376000, 37000,
979000, 17420000, 71065000, 46397000,
2005 Jan 06
0
Parametric Survival Models with Left Truncation, survreg
Hi,
I would like to fit parametric survival models to time-to-event data
that are left truncated. I have checked the help page for survreg and
looked in the R-help archive, and it appears that the R function survreg
from the survival library (version 2.16) should allow me to take account
of left truncation. However, when I try the command
2007 May 11
0
Tobit model and an error message
Dear R users:
I am using survreg for modeling left censored longitudinal data. When I am using the following code for fitting the tobit model I am getting some output with an warning message(highlighted with red color):
> survreg(Surv(y, y>=0, type='left')~x + frailty(id), cytokine.data, weight=w, dist='gaussian', scale=1)
Call:
survreg(formula = Surv(y, y >= 0, type
2012 Feb 10
0
coxme with frailty
A couple of clarifications for you.
1. I write mixed effects Cox models as exp(X beta + Z b), beta = fixed
effects coefficients and b = random effects coefficients. I'm using
notation that is common in linear mixed effects models (on purpose).
About 2/3 of the papers use exp(X beta)* c, i.e., pull the random
effects out of the exponent. Does it make a difference? Not much: b
will be
2013 Jun 12
2
survreg with measurement uncertainties
Hello,
I have some measurements that I am trying to fit a model to. I also
have uncertainties for these measurements. Some of the measurements
are not well detected, so I'd like to use a limit instead of the
actual measurement. (I am always dealing with upper limits, i.e. left
censored data.)
I have successfully run survreg using the combination of well detected
measurements and limits,