similar to: interpretation of coefficients in survreg AND obtaining the hazard function

Displaying 20 results from an estimated 200 matches similar to: "interpretation of coefficients in survreg AND obtaining the hazard function"

2010 Nov 13
2
interpretation of coefficients in survreg AND obtaining the hazard function for an individual given a set of predictors
Dear R help list, I am modeling some survival data with coxph and survreg (dist='weibull') using package survival. I have 2 problems: 1) I do not understand how to interpret the regression coefficients in the survreg output and it is not clear, for me, from ?survreg.objects how to. Here is an example of the codes that points out my problem: - data is stc1 - the factor is dichotomous
2010 Nov 16
1
Re : interpretation of coefficients in survreg AND obtaining the hazard function for an individual given a set of predictors
Thanks for sharing the questions and responses! Is it possible to appreciate how much the coefficients matter in one or the other model? Say, using Biau's example, using coxph, as.factor(grade2 == "high")TRUE gives hazard ratio 1.27 (rounded). As clinician I can grasp this HR as 27% relative increase. I can relate with other published results. With survreg the Weibull model gives a
2004 Nov 06
0
Re: Trusting and trusted domain (home mapping) problem
Hi Igor, Thanks so much for troubleshooting all this while and we found out none of our configuration is the problem but the source code. Hope that the samba team will modify to a working code so that I can deploy it. Actually my dateline to deploy is coming soon and I do not know what to do now..... when do you think the code will be modified and be released? Thanks so much for your help.
2012 Sep 10
1
Memory errors using lmer
Hello R List, I'm trying to run a linear mixed effects analysis on fairly large datasets with lmer (from the lme4 package) on a 32-bit Windows machine running XP with 3 GB of RAM. It's not working. (details below) I've researched the ff and bigmemory packages, but it appears they won't handle the mixed mode dataset I'm analyzing. It has some character fields for the
2007 Apr 18
0
[patch 2/9] Guest page hinting: unused / free pages on s390.
From: Martin Schwidefsky <schwidefsky@de.ibm.com> From: Hubertus Franke <frankeh@watson.ibm.com> From: Himanshu Raj <rhim@cc.gatech.edu> [patch 2/9] Guest page hinting: unused / free pages on s390. s390 uses the milli-coded ESSA instruction to set the page state. The page state is formed by four guest page states called block usage states and three host page states called block
2007 Apr 18
0
[patch 2/9] Guest page hinting: unused / free pages on s390.
From: Martin Schwidefsky <schwidefsky@de.ibm.com> From: Hubertus Franke <frankeh@watson.ibm.com> From: Himanshu Raj <rhim@cc.gatech.edu> [patch 2/9] Guest page hinting: unused / free pages on s390. s390 uses the milli-coded ESSA instruction to set the page state. The page state is formed by four guest page states called block usage states and three host page states called block
2004 Mar 11
3
widen the screen place
I have this problem: my screen seems not wide enough in R. My true computer monitor is very wide, so I wish that I can line up the text when I stretch it wide. i.e.: This is typical R returns: > summary(a) V1 V2 V3 V4 F:625 Min. :20020103 EMC : 34 Min. : 9300300 L:944 1st Qu.:20020530 BRW : 27 1st Qu.: 9323865
2011 Jan 24
1
Extracting information from text data
Hi R-Users,   Thanks in advance.   I am using R-2.12.0 on Windows XP.   I am trying to produce an n X m matrix from text data stored in different files. Where n = number of words (say w1, w2, …, wn). M is the number of documents (say d1, d2, …, dm)   A. Using package tm   I am using package tm to do the job. I have provided the code below:   > my.corpus <- Corpus(DirSource(my.path),
2011 Jul 11
3
NetCDF, open *.nc
*Good day!* I'm using Vista(x64) and R 2.13.0(x64) (with ncdf and RnetCDF packs) *I have a problem:* I'm trying to open **.nc* file and I use this script (http://r.789695.n4.nabble.com/Problem-downloading-and-opening-netcdf-file-td3046897.html#a3046965) library(ncdf) link <- "http://ibis.grdl.noaa.gov/SAT/SeaLevelRise/slr/slr_sla_gbl_free_all_66.nc" dest <-
2010 Dec 01
4
Sequence for repeated numbers
Hello fellows, I would like to create a sequence for repeated numbers in a dataset. For example: ID <- c(1:20) grade <- c(4,4,4,5,5,7,7,7,7,8,8,8,9,9,9,9,9,10,10,10) Data: ID Grade 1 4 2 4 3 4 4 5 5 5 6 7 7 7 8 7 9 7 (...) I would like to create a variable "sequence": Data: ID Grade Sequence: 1 4 1 2 4 2 3 4 3 4 5
2007 Feb 16
2
How is the line in av.plot calculated?
I suspect that the line in the added variable plot (library car) is a SLR of the residuals, but I can't seem to find this written anywhere. Can someone confirm this? Thanks. Tom [[alternative HTML version deleted]]
2016 Dec 22
2
struct bitfield regression between 3.6 and 3.9 (using -O0)
Here's our testcase: #include <stdio.h> struct flags { unsigned frog: 1; unsigned foo : 1; unsigned bar : 1; unsigned bat : 1; unsigned baz : 1; unsigned bam : 1; }; int main() { struct flags flags; flags.bar = 1; flags.foo = 1; if (flags.foo == 1) { printf("Pass\n"); return 0; } else {
2016 Dec 22
0
struct bitfield regression between 3.6 and 3.9 (using -O0)
On 12/21/2016 4:45 PM, Phil Tomson via llvm-dev wrote: > Here's our testcase: > > #include <stdio.h> > > struct flags { > unsigned frog: 1; > unsigned foo : 1; > unsigned bar : 1; > unsigned bat : 1; > unsigned baz : 1; > unsigned bam : 1; > }; > > int main() { > struct flags flags; > flags.bar = 1; >
2010 May 20
2
multiple 2 by 2 crosstabulations?
Hello, I have a dataframe (var_1, var_2, ..., var_n) and I would like to export summary statistics to Latex in the form of a table. I want specific summary statistics by crossing numerous variables 2x2 AT ONCE. In each cell I would like sometimes to have the median (Q1 - Q3), or frequency and proportion, etc. CrossTable, xtab, etc... do not allow for multiple 2 by 2 crosstabulation. The table
2008 Oct 15
0
Iterative estimation of linear regression model
Dear all I am intrested in making iterative estimation (thro' loop statements) of, say, linear regression model. For this purpose, I have written the following programme and that I have made use of a sample data (viz., exp.txt): ? Programme: ? # Linear regression modelling with sample data (try5.txt) # Repeated estimation through loop statement
2010 Nov 17
1
Problem downloading and opening netcdf file
I am trying to download and open an on-line netcdf file. I'm using Windows XP and R 2.11.1 Here's my script library(ncdf) link <- "http://ibis.grdl.noaa.gov/SAT/SeaLevelRise/slr/slr_sla_gbl_free_all_66.nc" dest <- "C:/temp/slr_sla_gbl_free_all_66.nc" download.file(url=link,destfile=dest) nc1 <- open.ncdf(dest) The file appears in my C:/temp
2004 Nov 28
1
Modifications to an abline
Dear List: I am working to generate graphs for individual students that will be created through a series of loops in Sweave. Before doing so, I am still trying to design the graph. The code for creating the barplot is below with some sample datapoints just made up for now. Ultimately, this chart will take data from an lme object using longitudinal student data. So, the dots represent the
2008 Aug 31
1
randomForest partial dependence plots
Hello ALL, I'm unsure how to interpret partial dependence plots and had 2 questions: Does the randomForest library scale partial dependence plots such that zero on the y axis is mean(y)? I am also unclear as to why the range of y on these plots is smaller than what is in the dataset. Also I have noticed that partial dependence plots cluster around the mean response. That is, if I'm
2007 Apr 30
3
syslinux booting from a third partion of ipod.
Hello, I hope that somebody can help me, I am trying to boot Knoppix from the third parition of an ipod nano with out loosing the player capabilities. After fdisk I have the following partitions: slr ~ # fdisk -l /dev/sda Disk /dev/sda: 4095 MB, 4095737344 bytes 255 heads, 63 sectors/track, 497 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Device Boot Start End
2016 Dec 23
2
struct bitfield regression between 3.6 and 3.9 (using -O0)
Given that this is compiled with -O0, would there a way to skip the Optimization of the Type-legalized selection DAG? It's fine until it optimizes the Type-legalized selection DAG into the Optimized Type-legalized selection DAG. Phil On Thu, Dec 22, 2016 at 10:29 AM, Friedman, Eli <efriedma at codeaurora.org> wrote: > On 12/21/2016 4:45 PM, Phil Tomson via llvm-dev wrote: > >