Displaying 20 results from an estimated 800 matches similar to: "Projection Pursuit"
2004 Mar 17
6
projection pursuit
Dear helpers
Does R have a package that performs projection pursuit density estimation? Or
anyone knows code in Matlab or C for example to do this?
Thank you all
Luis
2011 May 31
1
Projection Pursuit Index
Dear R-developers,
I am trying to experiment with projection pursuit (PP), and different
indexes for the same, especially using the tourr package. However, I've
noticed that a PP index in the said package is only a function of the
projected data. Could I modify the function so that the index sees the
projection bases instead (or in addition to the data) ?
Alternatively, if some other package
2008 Oct 28
1
Source code for ppr (Projection Pursuit Regression)
Dear R users,
I am looking for the source code of the implementation of ppr (Projection
Pursuit Regression) in R.
It will be great if citations of the source papers on which the
implementation is based, are also provided.
Thank you,
Arvind Iyer,
Grad student, Deptt. of Biomedical Engineering
Viterbi School of Engineering
University of Southern California, Los Angeles
[[alternative HTML
2000 Sep 01
1
Help with Projection Pursuit, ppr().
Hi,
Recently, I installed the 1.1.0 version of R (for Windows), since it includes an implementation of Projection Pursuit (I failed to write my own version of PP as a standalone C++ program).
As far as I know, R offers two interfaces/sintax for the ppr() function. The first one requieres a regression formula and a data frame. The other requieres X, a matrix with the explanatory variables, and Y,
2006 Apr 10
6
"Pursuit of Happiness" ? Are you sure ?
I always kind of liked this title (the title of Davids super Rails
presentation), but then I saw this quote today:
"The pursuit of happiness is a most ridiculous phrase; if you pursue
happiness you''ll never find it." by C.P Snow
I''d have to agree with this quote. It''s the same as love... go looking for
it, and you''ll never find it.
Soooo... David,
2011 Feb 16
0
Constraints in projection pursuit regression
Hi,
I am solving a projection pursuit regression problem, of the
form y = \sum_i f_i (a_i^T x), where a_i are unknown directions, while
f_i are unknown univariate link functions. The following is known about
each f_i:
1. f_i (0) = 0 (that is, each f_i passes through the origin)
2. f_i is monotonic.
Is there a way to ensure that the function ppr() in R produces solutions that respect the
2005 May 13
0
df and gcvpen for parameters selection on projection pursuit regression
Hello,
I am using projection pursuit regression parameters selection.
Does anyone has experience on the range to test for df parameter (spline kernel) and gcvpen (gcvspline kernel)?
I don't find any information about this.
Thanks in advance.
Joao Moreira
[[alternative HTML version deleted]]
2010 Jul 29
1
Crash report: projection pursuit & predict
Folks,
The projection pursuit regression function in the base R seems to crash when the optimization level is set to zero, i.e. the initial ridge terms are accepted without refitting. I encountered this problem in an out-of-sample prediction exercise using predict. But further investigation suggests the issue is with the ppr fit and predict just sppeds up the crash. The other optlevels seem to be
2006 Jan 23
2
Master's project to coerce linux nvidia drivers to run generalised linear models
Hi,
I am working with a friend on a master's project. Our laboratory does a
lot of statistical analysis using the R stats package and we also have a
lot of under-utilised nvidia cards sitting in the back of our networked
linux machines. Our idea is to coerce the linux nvidia driver to run
some of our statistical analysis for us. Our first thought was to
specifically code up a version of
2009 Jan 07
0
fixed effect significance_NB mixed models_further pursuit
7 Jan 09
Hello,
I am using R version 2.7.0 in a Windows XP context.
I am also using the glmm.admb package (created by Dave Fournier, Hans
Skaug, and Anders Nielson) to run mixed-effects negative binomial
models.
To the best of my knowledge and ability, I have searched and studied
the R-help, R-sig-mixed models, and ADMB NBMM for R (through Otter
Research Ltd) list servs; R help
2005 Nov 09
1
Order of terms in a model specification...
Hi,
Sorry for this one as its pretty basic but I've taken a look for info and
couldn't find any...
My question is, does the order of main effect terms in a model specification
have any impact on the model R fits or not. (in particular when using lm).
ie
Can A~X+Y+Z lead to different results to A~Z+Y+X, and if so in what
circumstances, and how much should I worry about it?
I believe
2010 Apr 14
3
pdMat
Alguien tiene experiencia en escribir una pdMat. Para aquellos que no lo
recuerden son las matrices de covarianzas de los efectos aleatorios que
ajusta la función lme de la librería nlme
Estas matrices tiene especial importancia en aplicaciones de genética de
poblaciones y en particular en mapeo de asociación. Pinheiro y Bates dicen
que el usuario puede crear sus propias pdMat y sugiere como
2005 Apr 11
1
multi-class modeling
Hi,
Just wonder if someone could comment on using linear
discriminant analysis (LDA) vs. multinomial logistic
regression in multi-class classification/prediction
(nomial dependent variable, not ordinal)? What kind of
difference in results can I expect from the 2 methods,
which is better or more appropriate, or under what
condiditon should I used one instead of the other? And
is there other
2005 May 04
1
Difference between "tree" and "rpart"
In the help for rpart it says, "This differs from the tree function
mainly in its handling of surrogate variables." And it says that an
rpart object is a superset of a tree object. Both cite Brieman et al.
1984. Both call external code which looks like martian poetry to me.
I've seen posts in the archives where BDR, and other knowledgeable
folks, have said that rpart() is to be
2006 Jan 18
1
Canonical Variance Analysis by any other name?
I've been asked about "Canonical Variance Analysis" (CVA). I don't
see any reference to it searching the R site. Does it go by other
names?
Genstat describes it thus:
Canonical variates analysis operates on a within-group sums of squares
and products matrix, calculated from a set of variates and factor that
specifies the grouping of units. It finds linear combinations of the
2000 Mar 20
1
CART and the `tree' contrib package
Dear R people,
I was recently reading the book `Classification and Regression Trees' by
Breiman. This book talks about the CART program. Both Splus and R have
implementations of this. However, the book talks about the possibility of
extending the existing `standard' set of questions (for continuous
variables, these are of the form X < c where X is the variable, c some
const) to
2017 Apr 15
2
help on listener connection stats
Hi,
I see two statistics on the admin web page.
On my system "listeners" is usually 0. Sometimes I will
actually catch it at 1 or 2, but it is 0 99% of the time.
On the other hand "listener_connections" is always
going up. If I refresh the screen it goes up by a few.
If I wait an hour and refresh it goes up by tens.
Over a few days it goes up by hundreds.
What is this
2000 Mar 09
1
What is XGobi and what is it for?
I've seen XGobi come up in here from time to time, and I wonder what is
it, what is it for, do I need it? How about throwing me a bone here!
--
Paul E. Johnson email: pauljohn at ukans.edu
Dept. of Political Science http://lark.cc.ukans.edu/~pauljohn
University of Kansas Office: (785) 864-9086
Lawrence, Kansas 66045 FAX: (785)
2007 Dec 20
2
factor manipulation: edgelist to a matrix?
Hello All,
I have had considerable bad luck with attempting the following with for
loops. Here is the problem:
# Suppose we have a data.frame with the following data, which can be
considered a type of edgelist (for those with networks backgrounds):
#
# V1 V2
# 1 A
# 1 A
# 1 B
# 2 A
# 3 C
# 3 A
# 3 C
# 3 B
#
# I want the output of the function to produce a matrix, such that #each
factor of
2005 Jul 01
1
p-values for classification
Dear All,
I'm classifying some data with various methods (binary classification). I'm interpreting the results via a confusion matrix from which I calculate the sensitifity and the fdr. The classifiers are trained on 575 data points and my test set has 50 data points.
I'd like to calculate p-values for obtaining <=fdr and >=sensitifity for each classifier. I was thinking about