Displaying 20 results from an estimated 120 matches similar to: "newbie ifelse matrix question"
2008 Feb 15
1
Questions about EM algorithm
Dear all:
Assume I have 3 distributions, x1, x2, and x3.
x1 ~ normal(mu1, sd1)
x2 ~ normal(mu2, sd2)
x3 ~ normal(mu3, sd3)
y1 = x1 + x2
y2 = x1 + x3
Now that the data I can observed is only y1 and y2. It is
easy to estimate (mu1+m2), (mu1+mu3), (sd1^2+sd2^2) and
(sd1^2+sd3^2) by EM algorithm since
y1 ~ normal(mu1+mu2, sqrt(sd1^2+sd2^2)) and
y2 ~ normal(mu1+mu3, sqrt(sd1^2+sd3^2))
However, I want
2011 Jan 10
2
Calculating Portfolio Standard deviation
Dear R helpers
I have following data
stocks <- c("ABC", "DEF", "GHI", "JKL")
prices_df <- data.frame(ABC = c(17,24,15,22,16,22,17,22,15,19),
DEF = c(22,28,20,20,28,26,29,18,24,21),
GHI = c(32,27,32,36,37,37,34,23,25,32),
2010 May 18
25
Very serious performance degradation
Hi,
I''m running Opensolaris 2009.06, and I''m facing a serious performance loss with ZFS ! It''s a raidz1 pool, made of 4 x 1TB SATA disks :
zfs_raid ONLINE 0 0 0
raidz1 ONLINE 0 0 0
c7t2d0 ONLINE 0 0 0
c7t3d0 ONLINE 0 0 0
c7t4d0 ONLINE 0 0
2012 Jun 04
1
simulation of modified bartlett's test
Hi, I run this code to get the power of the test for modified bartlett's
test..but I'm not really sure that my coding is right..
#normal distribution unequal variance
asim<-5000
pv<-rep(NA,asim)
for(i in 1:asim)
{print(i)
set.seed(i)
n1<-20
n2<-20
n3<-20
mu<-0
sd1<-sqrt(25)
sd2<-sqrt(50)
sd3<-sqrt(100)
g1<-rnorm(n1,mu,sd1)
g2<-rnorm(n2,mu,sd2)
2010 Jul 18
2
loop troubles
Hi all, I appreciate the help this list has given me before. I have a
question which has been perplexing me. I have been working on doing a
Bayesian calculating inserting studies sequentially after using a
non-informative prior to get a meta-analysis type result. I created a
function using three iterations of this, my code is below. I insert prior
mean and precision (I add precision manually
2006 Nov 24
1
barplot help needed
hello,
I would like to create the following barplot:
I have 4 different data sets (same length + stddev for each data point)
data1
sd1
data2
sd2
data3
sd3
data4
sd4
now, I'd like to plot in the following way:
data1[1],data2[1],data3[1],data4[1] with it's sd-values side-by-side at
one x-axis label (named "position 1") and each bar in different colors.
2011 Nov 11
1
Fwd: Use of R for VECM
----- Forwarded Message -----
From: vramaiah at neo.tamu.edu
To: "bernhard pfaff" <bernhard.pfaff at pfaffikus.de>
Sent: Friday, November 11, 2011 9:03:11 AM GMT -06:00 US/Canada Central
Subject: Use of R for VECM
Hello Fellow R'ers
I am a new user of R and I am applying it for solving Bi-Variate (Consumption and Output) VECM with Co-Integration (I(1)) with three lags on
2007 Nov 17
11
slog tests on read throughput exhaustion (NFS)
I have historically noticed that in ZFS, when ever there is a heavy
writer to a pool via NFS, the reads can held back (basically paused).
An example is a RAID10 pool of 6 disks, whereby a directory of files
including some large 100+MB in size being written can cause other
clients over NFS to pause for seconds (5-30 or so). This on B70 bits.
I''ve gotten used to this behavior over NFS, but
2008 Aug 22
1
sd(NA)
Hi All:
This was discussed in the R-developers list (see the thread
starting at http://tolstoy.newcastle.edu.au/R/e3/devel/
07/12/0560.html). It has to do with the behavior of sd() when the
entire vector is NA. The behavior has changed between 2.6 and 2.7.1
as follows:
Run in Version 2.7.1
> tt<-rep(NA, 10)
> mean(tt, na.rm=T)
[1] NaN
>sd(tt, na.rm=T)
Error in var(x,
2011 Nov 29
1
Create an identifier variable
I am needing to create a new identifier variable for a data set which has no
ID variable in the original file. I am basically wanting to take the count
of each row and add "sd" in front of it - so it would look like so
sd1
sd2
sd3
sd4
sd5 etc.......
I have no idea how to do this. I am a SAS user trying to learn R. This
question may have been answered previously, but I could be
2008 Jan 10
2
NCQ
fun example that shows NCQ lowers wait and %w, but doesn''t have
much impact on final speed. [scrubbing, devs reordered for clarity]
extended device statistics
device r/s w/s kr/s kw/s wait actv svc_t %w %b
sd2 454.7 0.0 47168.0 0.0 0.0 5.7 12.6 0 74
sd4 440.7 0.0 45825.9 0.0 0.0 5.5 12.4 0 78
sd6 445.7 0.0
2011 Apr 29
1
question of VECM restricted regression
Dear Colleague
I am trying to figure out how to use R to do OLS restricted VECM regression. However, there are some notation I cannot understand.
Please tell me what is 'ect', 'sd' and 'LRM.dl1 in the following practice:
#OLS retricted VECM regression
data(denmark)
sjd <- denmark[, c("LRM", "LRY", "IBO", "IDE")]
sjd.vecm<-
2012 Jun 14
0
fixed trimmed mean for j-group
Hello...i want to find the empirical rate for type 1 error using fixed
trimmed mean. To make it easy, i'm referring to journal given by this
website
http://www.academicjournals.org/ajmcsr/PDF/pdf2011/Yusof%20et%20al.pdf.
I already run the programme and there is no error in it but i got zero for
the empirical rate of type 1 error. The empirical rate for the type 1 error
given in the journal
2014 Dec 11
5
[LLVMdev] dynamic data dependence extraction using llvm
Hi LLVM-ers,
I try to develop my custom dynamic data dependence tool (focusing on nested
loops), currently I can successfully get the trace including load/store
address, loop information, etc.
However, when I try to analyze dynamic data dependence based on the
pairwise method described in [1], the load/store for iteration variables
may interfere my analysis (I only care about the load/store for
2007 Apr 30
4
B62 AHCI and ZFS
Hardware Supermicro X7DAE (AHCI BIOS) dual Intel Woodcrest processors, 6 x Western Digital Raptor SATA drives.
I have installed b62 running 64 bit succesfully on a PATA drive. The BIOS is configured to access the SATA drives in native mode using hte AHCI Bios.
I have 6 SATA II drives accessed via the Solaris AHCI driver. I have created a ZFS file system across all 6 drives. This works fine until
2012 Jul 07
0
fixed trimmed mean for group
Hello,
I haven't found errors in your code. I implemented the test in the paper
(the first, fixed symetric mean) and it also gives me zero Type I
errors, when alpha = 0.05. Try to see the value of min(pv) or to plot
the histogram of 'pv', hist(pv) and you'll see that there are no
significant p-values, at that level.
Anyway I'll continue to look at it, but my first
2014 Dec 11
2
[LLVMdev] dynamic data dependence extraction using llvm
Dear Dibyendu,
Thanks for your response. :-)
> If you are looking for only dependences which are inter-iteration
(dependence distance != 0 ) you can do a post-pass on the ld/st addresses
collected
Yes, I am more interested in inter-iteration dependence. Could you
provide more information or some links on post-pass approach? I have no
idea on your method. :-)
> eliminate such
2009 Jul 13
7
OpenSolaris 2008.11 - resilver still restarting
Just look at this. I thought all the restarting resilver bugs were fixed, but it looks like something odd is still happening at the start:
Status immediately after starting resilver:
# zpool status
pool: rc-pool
state: DEGRADED
status: One or more devices has experienced an unrecoverable error. An
attempt was made to correct the error. Applications are unaffected.
action: Determine
2010 Feb 12
13
SSD and ZFS
Hi all,
just after sending a message to sunmanagers I realized that my question
should rather have gone here. So sunmanagers please excus ethe double
post:
I have inherited a X4140 (8 SAS slots) and have just setup the system
with Solaris 10 09. I first setup the system on a mirrored pool over
the first two disks
pool: rpool
state: ONLINE
scrub: none requested
config:
NAME
2014 Dec 12
2
[LLVMdev] dynamic data dependence extraction using llvm
Dear Dibyendu and Mobi,
Thanks for your help! :-)
I finally figure it out. The solution is really simple. I just need to
generate a new bitcode file with the following command:
-----
opt -mem2reg -indvars test1.bc -o test2.bc
-----
Then the load/store for induction variables will be removed and replaced by
PHI instructions and all remaining load/store instructions are those I am
interested in. I