Displaying 20 results from an estimated 3000 matches similar to: "computing misclassification table for tree objects"
2010 Nov 22
1
using rpart with a tree misclassification condition
Hello
I want to build a classification tree for a binary response variable
while the condition for the final tree should be :
The total misclassification for each group (zero or one) will be less then
10% .
for example: if I have in the root 100 observations, 90 from group 0 and 10
from group 1, I want that in the final tree a maximum of 9 and 1
observations out of group 0 and 1, respectively,
2002 Jan 07
1
is then an equivalent of partition.tree for rpart?
partition.tree plots in 2d the partition of a classification tree produced by the function
tree (assuming the data frame from which it was computed has two continuous predictors).
I get an error when I feed a tree produced by rpart to partition.tree (since trees
produced by rpart are superclasses of those produced by tree). Is there an equivalent of
partition.tree for objects of class rpart?
2011 Sep 02
2
misclassification rate
Hi users
I'm student who is struggling with basic R programming. Would you please
help me with this problem.
"My english is bad" I hope that my question is clear:
I have a matrix in wich there are two colmns( yp, yt)
Yp: predicted values from my model.
yt: true values ( my dependante variable y is a categorical;3 modalities
(0,1,2)
I don't know how to procede to calculate the
2002 Mar 05
1
no labels when plotting dendrograms
I'd like to be able to cut dendrograms at a height I specify
and then plot the resulting subtrees. I wanted to use the
dendrogram object for this purpose because there doesn't seem
to be a canned way to cut a hclust object and get a list of
hclust objects, but there is a function (cut) that does that
for dendrograms. The problem I'm having is that when I plot
a dendrogram, I
2002 Jan 13
1
changing the ordering of leaves in a dendrogram
I'd like to change the way plot.hclust displays an hclust object. Here's
a description of how it's done now, from the R documentation of hclust:
In hierarchical cluster displays, a decision is needed at each merge
to specify which subtree should go on the left and which on the right.
Since, for n observations there are n-1 merges, there are 2^{(n-1)}
possible
2007 Jun 16
0
Function for misclassification rate/type I,II error??
HI
Is there any function in R that tells us error rate(misclassification rate)
for logistic regression type classification?
i also want to know the function to determine type I and type II error.
I have found a link where "misclass" and "confusion" are used. But I dont
know the package name.
http://alumni.media.mit.edu/~tpminka/courses/36-350.2001/lectures/day32/
2002 Mar 26
2
does function predplot still exist?
Venables and Ripley's MASS 3ed mentions a function predplot, used for
plotting results of lda and qda. But even with the library MASS loaded,
R (1.4.1) tells me predplot doesn't exist, and when I search the 1.4.1
documentation for predplot, it finds nothing. Does predplot still
exist somewhere, or has it been replaced by a new and improved version
with a different name? Thanks.
2001 Dec 29
1
load in 1.4.0 not working for me
I'm having trouble with save and load in R 1.4.0 (on RH 7.2) I originally
noticed this with a large R object, but it happens with a tiny one as well:
> a <- c(1,2,3)
> save(a,file="test.RData")
> a1 <- load("test.RData")
> print(a1)
NULL
The file "test.RData" is created, and it does have something in it. When
I tried saving a
2009 May 12
1
questions on rpart (tree changes when rearrange the order of covariates?!)
Greetings,
I am using rpart for classification with "class" method. The test data is
the Indian diabetes data from package mlbench.
I fitted a classification tree firstly using the original data, and then
exchanged the order of Body mass and Plasma glucose which are the
strongest/important variables in the growing phase. The second tree is a
little different from the first one. The
2010 Apr 30
1
how is xerror calculated in rpart?
Hi,
I've searched online, in a few books, and in the archives, but haven't seen
this. I believe that xerror is scaled to rel error on the first split.
After fitting an rpart object, is it possible with a little math to
determine the percentage of true classifications represented by a xerror
value? -seth
--
View this message in context:
2008 Feb 24
1
what missed ----- CART
Hi all,
Can anyone who is familar with CART tell me what I missed in my tree code?
library (MASS)
myfit <- tree (y ~ x1 + x2 + x3 + x4 )
# tree.screens () # useless
plot(myfit); text (myfit, all= TRUE, cex=0.5, pretty=0)
# tile.tree (myfit, fgl$type) # useless
# close.screen (all= TRUE) # useless
My current tree plot resulted from above code shows as:
2011 May 12
1
Saving misclassified records into dataframe within a loop
Greetings R world,
I know some version of the this question has been asked before, but i need
to save the output of a loop into a data frame to eventually be written to a
postgres data base with dbWriteTable. Some background. I have developed
classifications models to help identify problem accounts. The logic is this,
if the model classifies the record as including variable X and it turns out
2009 Mar 11
2
Couple of Questions about Classification trees
So I have 2 sets of data - a training data set and a test data set. I've been
doing the analysis on the training data set and then using predict and
feeding the test data through that. There are 114 rows in the training data
and 117 in the test data and 1024 columns in both. It's actually the same
set of data split into two. The rows are made of 5 different numbers. They
do represent
2011 Oct 25
2
Logistic Regression - Variable Selection Methods With Prediction
Hello,
I am pretty new to R, I have always used SAS and SAS products. My
target variable is binary ('Y' and 'N') and i have about 14 predictor
variables. My goal is to compare different variable selection methods
like Forward, Backward, All possible subsests. I am using
misclassification rate to pick the winner method.
This is what i have as of now,
Reg <- glm (Graduation ~.,
2005 Oct 14
1
Predicting classification error from rpart
Hi,
I think I'm missing something very obvious, but I am missing it, so I
would be very grateful for help. I'm using rpart to analyse data on
skull base morphology, essentially predicting sex from one or several
skull base measurements. The sex of the people whose skulls are being
studied is known, and lives as a factor (M,F) in the data. I want to
get back predictions of gender, and
2012 Aug 19
1
e1071 - tuning is not giving the best within the range
Hi everybody,
I am new in e1071 and with SVMs. I am trying to understand the performance
of SVMs but I face with a situation that I thought as not meaningful.
I added the R code for you to see what I have done.
/set.seed(1234)
data <- data.frame( rbind(matrix(rnorm(1500, mean = 10, sd = 5),ncol = 10),
matrix(rnorm(1500, mean = 5, sd = 5),ncol = 10)))
class <- as.factor(rep(1:2,
2007 Jan 29
3
comparing random forests and classification trees
Hi,
I have done an analysis using 'rpart' to construct a Classification Tree. I
am wanting to retain the output in tree form so that it is easily
interpretable. However, I am wanting to compare the 'accuracy' of the tree
to a Random Forest to estimate how much predictive ability is lost by using
one simple tree. My understanding is that the error automatically displayed
by the two
2009 Apr 27
1
question about adaboost.
Hello,
I would like to know how to obtain the misclassification error when performing a boosting analisis with ADABAG package?
With:
> prop.table(Tesis.boostcv$confusion)
I obtain the confusion matrix, but not the overall missclassification error.
Thanks in advance,
BSc. Cecilia Lezama
Facultad de Ciencias - UDELAR
Montevideo - Uruguay.
[[alternative HTML version deleted]]
2007 Jun 12
3
Appropriate regression model for categorical variables
Dear users,
In my psychometric test i have applied logistic regression on my data. My
data consists of 50 predictors (22 continuous and 28 categorical) plus a
binary response.
Using glm(), stepAIC() i didn't get satisfactory result as misclassification
rate is too high. I think categorical variables are responsible for this
debacle. Some of them have more than 6 level (one has 10 level).
2008 May 21
1
How to use classwt parameter option in RandomForest
Hi,
I am trying to model a dataset with the response variable Y, which has
6 levels { Great, Greater, Greatest, Weak, Weaker, Weakest}, and
predictor variables X, with continuous and factor variables using
random forests in R. The variable Y acts like an ordinal variable, but
I recoded it as factor variable.
I ran a simulation and got OOB estimate of error rate 60%. I validated
against some