Displaying 20 results from an estimated 35 matches for "0.0625".
Did you mean:
0.0025
2013 Feb 21
2
ggplot2, geomtile fill assignment
Dear R help,
I have some readings in three dimensions (x, y, z) and an amplitude for
each. I'd like to visualize the data using ggplot, using tile plots, as I
have some additional point data I would like to eventually overlay on the
tile plots.
I would like to subset the data by sections, slices if you will, in the z
dimension, and plot the data for that slice.
I can do all of this, but am
2012 Oct 15
3
Referring to matrix elements by name, iteratively
#Here is a vector of IDs
> cwaves
[1] 86 90 185 196 197 209 210 215 216 217 218
#Here is a matrix. The rows and columns correspond to the IDs in cwaves, and
the matrix is populated with a coefficient
> mat
86 90 185 196 209 210 215 216 217 218
86 0 0 0 0 0 0.00000 0 0.000000 0.000000 0.000000
90 0 0 0 0 0 0.00000 0
2009 Mar 24
3
confidence interval or error of x intercept of a linear regression
Hello all,
This is something that I am sure has a really suave solution in R, but I can't quite figure out the best (or even a basic) way to do it.
I have a simple linear regression that is fit with lm for which I would like to estimate the x intercept with some measure of error around it (confidence interval). In biology, there is the concept of a developmental zero - a temperature under
2005 May 04
1
Calculate median from counts and values
I am tangled with a syntax question. I want to calculate basic statistics
for a large dataset provided in weights and values and I can't figure out
an elegant way to expand the data.
For example here are the counts:
> counts
n4 n3 n2 n1 p0 p1 p2 p3 p4
1 0 0 0 1 1 3 16 55 24
2 0 0 0 0 2 8 28 47 15
3 1 17 17 13 4 5 12 24 8
...
and the values:
> values
2019 May 06
2
R optim(method="L-BFGS-B"): unexpected behavior when working with parent environments
Optim's Nelder-Mead works correctly for this example.
> optim(par=10, fn=fn, method="Nelder-Mead")
x=10, ret=100.02 (memory)
x=11, ret=121 (calculate)
x=9, ret=81 (calculate)
x=8, ret=64 (calculate)
x=6, ret=36 (calculate)
x=4, ret=16 (calculate)
x=0, ret=0 (calculate)
x=-4, ret=16 (calculate)
x=-4, ret=16 (memory)
x=2, ret=4 (calculate)
x=-2, ret=4 (calculate)
x=1, ret=1
2009 Jun 25
2
stringsAsFactors has no impact in expand.grid()?
Hi
I have the feeling, that the argument stringsAsFactors has no impact in the
function expand.grid:
a <- c("PR", "NC", "A2", "BS")
b <- c(1, 0.5, 0.25, 0.125, 0.0625, 0.03125)
class(expand.grid(css, fscs, stringsAsFactors=FALSE)[[1]])
[1] "factor"
class(expand.grid(css, fscs, stringsAsFactors=TRUE)[[1]])
[1] "factor"
Also, when
2011 Jan 20
1
Problems with ecodist
Dear Dr.Goslee and anyone may intrested in matrix manipulate,
I am using your ecodist to do mantel and partial mantel test, I have
locality data and shape variation data, and the two distance matrixs are
given as belowings. When I run the analysis, it is always report that the
matrix is not square, but I didn't know what's wrong with my data. Would you
please help me on this. I am quite
2014 Feb 02
5
[LLVMdev] [RFC] BlockFrequency is the wrong metric; we need a new one
Right now, all profile information is funneled through two analysis passes
prior to any part of the optimizer using it.
First, we have BranchProbabilityInfo, which provides a simple interface to
the simplest form of profile information: local and relative branch
probabilities. These merely express the likelihood of taking one of a
mutually exclusive set of exit paths from a basic block. They are
2011 Jul 06
1
Group Data indexed by n Variables
Hello,
the more general thing I'd like to learn here is how to compute Function of
Data on the basis of grouping determiend by n variables.
In terms of the reason why I am interested in this, I need to compute the
average of my data based on the value of the month and day across years. I
have come up withy the code below which, as far as I can see, does what I
need but getting either a more
2019 May 06
0
R optim(method="L-BFGS-B"): unexpected behavior when working with parent environments
That's consistent/not surprising if the problem lies in the numerical
gradient calculation step ...
On 2019-05-06 10:06 a.m., Ravi Varadhan wrote:
> Optim's Nelder-Mead works correctly for this example.
>
>
>> optim(par=10, fn=fn, method="Nelder-Mead")
> x=10, ret=100.02 (memory)
> x=11, ret=121 (calculate)
> x=9, ret=81 (calculate)
> x=8, ret=64
2011 Jul 04
3
loop in optim
Hi
May you help me correct my loop function.
I want optim to estimates al_j; au_j; sigma_j; b_j by looking at 0 to 20,
21 to 40, 41 to 60 data points.
The final result should have 4 columns of each of the estimates AND 4 rows
of each of 0 to 20, 21 to 40, 41 to 60.
###MY code is
n=20
runs=4
out=matrix(0,nrow=runs)
llik = function(x)
{
al_j=x[1]; au_j=x[2]; sigma_j=x[3]; b_j=x[4]
2011 Jul 14
1
plotting x y z data from an irregular grid
Hi,
I'm trying to plot some data (z) that is linked to lat&long coordinates
(x&y). These coordinates are not however on a regular grid. I also have some
shapefiles on which I would like to overlay the data. I can plot the
shapefiles (country/city outlines) and overplot the data, but only using
quilt.plot because I otherwise always get the error message that
'Error in
2019 May 06
1
R optim(method="L-BFGS-B"): unexpected behavior when working with parent environments
It seems that it's an old bug that was found in some other packages, but
at that time not optim:
https://bugs.r-project.org/bugzilla/show_bug.cgi?id=15958
and that Duncan Murdoch posted a patch already last Friday :)
Thomas
Am 06.05.2019 um 16:40 schrieb Ben Bolker:
> That's consistent/not surprising if the problem lies in the numerical
> gradient calculation step ...
>
2010 May 10
1
how to get p-value from ave
Hi there,
I checked google for aov. usually one uses summary to see whether the p-value is small.
but I want to put aov in my script. how can I get the p-value, (0.1115, 0.6665, 0.6665 in the following example)?
thanks
YU
> datafilename="http://personality-project.org/r/datasets/R.appendix2.data"
> data.example2=read.table(datafilename,header=T)
> aov.ex2 =
2014 Feb 03
4
[LLVMdev] [RFC] BlockFrequency is the wrong metric; we need a new one
On Sun, Feb 2, 2014 at 6:18 PM, Andrew Trick <atrick at apple.com> wrote:
>
> On Feb 2, 2014, at 2:13 AM, Chandler Carruth <chandlerc at gmail.com> wrote:
>
> > Right now, all profile information is funneled through two analysis
> passes prior to any part of the optimizer using it.
> >
> > First, we have BranchProbabilityInfo, which provides a simple
2008 Apr 03
3
summary(object, test=c("Roy", "Wilks", "Pillai", ....) AND ellipse(object, center=....)
Dear All,
I would be very appreciative of your help with the following
1). I am running multivariate multiple regression through the manova() function (kindly suggested by Professor Venables) and getting two different answers for test=c("Wilks","Roy","Pillai") and tests=c("Wilks","Roy",'"Pillai") as shown below. In the
2005 Dec 10
2
quantile regression problem
Dear List members,
I would like to ask for advise on quantile regression in R.
I am trying to perform an analysis of a relationship between species abundance
and its habitat requirements -
the habitat requirements are, however, codes - 0,1,2,3... where 0<1<2<3
and the scale is linear - so I would be happy to treat them as continuos
The analysis of the data somehow does not work, I am
2010 Aug 10
1
one (small) sample wilcox.test confidence intervals
Dear R people,
I notice that the confidence intervals of a very small sample (e.g. n=6) derived from the one-sample wilcox.test are just the maximum and minimum values of the sample. This only occurs when the required confidence level is higher than 0.93. Example:
> sample <- c(1.22, 0.89, 1.14, 0.98, 1.37, 1.06)
> summary(sample)
Min. 1st Qu. Median Mean 3rd Qu. Max.
2014 Feb 05
4
[LLVMdev] [RFC] BlockFrequency is the wrong metric; we need a new one
On Feb 3, 2014, at 12:13 AM, Andrew Trick <atrick at apple.com> wrote:
> On Feb 2, 2014, at 6:55 PM, Chandler Carruth <chandlerc at gmail.com> wrote:
>
>> On Sun, Feb 2, 2014 at 6:18 PM, Andrew Trick <atrick at apple.com> wrote:
>>
>> On Feb 2, 2014, at 2:13 AM, Chandler Carruth <chandlerc at gmail.com> wrote:
>>
>> > Right now,
2019 May 03
2
R optim(method="L-BFGS-B"): unexpected behavior when working with parent environments
Yes, I think you are right. I was at first confused by the fact that after the optim() call,
> environment(fn)$xx
[1] 10
> environment(fn)$ret
[1] 100.02
so not 9.999, but this could come from x being assigned the final value without calling fn.
-pd
> On 3 May 2019, at 11:58 , Duncan Murdoch <murdoch.duncan at gmail.com> wrote:
>
> Your results below make it look like a