Displaying 20 results from an estimated 1000 matches similar to: "remove descriptions from output"
2012 Sep 14
4
concatenating two vectors
Dear all,
I want to concatenate the elements of two vectors such as
a<-c("a1","a2")
b<-c("b1","b2")
and obtain
"a1b1", "a1b2","a2b1","a2b2"
I tried the paste and paste0 functions, but they yielded elementwise
concatenation such as
"a1b1","a2b2"
I am wondering that is there an efficient
2012 Sep 16
4
two questions about character manipulation
Dear all,
I want to manipulate a character string such as
ex<-"cbind(data$response1,data$response2)"
in R in two ways:
1) extracting the "response1" portion of ex
2) replacing "$" with "."
I am wondering that is it possible efficiently doing these in R?
Best
Ozgur
--
View this message in context:
2012 Mar 03
3
submission of some functions to base package
Dear all,
Actually I could not decide who to contact, then decided to post here. If
not appropriate sorry for that.
I have written some functions in R which might be supposed to available in
the base package but not available. I am wondering that how can I submit
those functions to that package?
Best
Ozgur
-----
************************************
Ozgur ASAR
Research Assistant
Middle East
2012 May 31
2
Loop question
Hello,
I have a dataframe (Lx) with 5 Lb, and 5 Lw variables. I want to
create several variables according to the loop presented below (data
attached).
Lx <- read.csv("Lx.csv", header=T, sep=",")
for (i in 1:20) {
Lx$sb1[i] <- Lx$Lb1[i+1]/Lx$Lb1[i]
Lx$sb2[i] <- Lx$Lb2[i+1]/Lx$Lb2[i]
Lx$sb3[i] <- Lx$Lb3[i+1]/Lx$Lb3[i]
Lx$sb4[i] <-
2012 May 31
1
Warning message: numerical expression has 1000 elements: only the first used
Hi,
Your mistake seems to be in
sum(v[1:x])
You create "x" as a vector but your treat it as a single number.
v[1:x] expects "x" to be a single number and only considers its first
element which is 1.
If I understand your query correctly, the following might handle your
problem:
sum.vec <-NULL
for (x in 1:1000){
t <- rbinom(1000, 1, 0.5)
v <- replace(t,t==0,-1)
2012 May 27
7
Customized R Regression Output?
Hello R-Experts,
I am facing the problem that I have to estimate several parameters for a lot
of different dependent variables.
One single regression looks something like this:
y = beta0 + beta1 * x1 + beta2 * x2 + beta3 * x1 * x2 + beta4 * x4 + beta5 *
lag(x4,-1)
where y is the dependent variable and xi are the independent ones. Important
to me are the different estimates of betai and their
2012 Jun 28
3
Error: could not find function
Hi
Do I have to load/download the whole package over and over again everytime
when I use the R-Project?
It often prompt me the error message "could not find function"- but I'm
rather sure I have used the function before so I'm not entirely sure it
couldn't find it again when I shut down the computer or so.
Is it to do with directory or clashing with other software?
Any
2012 May 23
2
Special characters in an R package manual
Dear all,
I have some trouble with special characters while building my R package. I
tried to follow the usual LATEX format, but could not fix the problem:
For instance, for "greater than or equal", I tried "\geq", but R says that
this is an unknown macro.
Could anyone direct me how to solve this issue?
Best
Ozgur
-----
************************************
Ozgur ASAR
2010 Dec 13
2
help on SAS Macro in R
Dear Researchers,
I am looking for to read a SAS macro in R. Although I searched it on web, I couldn’t find anything.
Can you help me or direct me?
Thank you for your interest and patience.
Best.
Ozgur
[[alternative HTML version deleted]]
2012 May 26
3
How to measure level of similarity of two data frames
Hi group,
I've been thinking of calculating euclidean distance between each column of
a data frames that each consists of standardized numerical columns.
However, I don't know if there's a way of summarizing the overall distance
by some kind of metrics. If anyone know a proper way of doing so and/or a
package I would greatly appreciate your suggestions. Thanks very much!
Kel
--
2012 May 28
1
simulation of levene's test
hello,
I try to run simulation of levene's test to find the p-value but the error
of replacement has length zero occur, could anyone help me to fix this
problem?
asim <- 1000
pv<-rep(NA,asim)
for(i in 1:asim)
{print(i)
set.seed(i)
g1 <- rnorm(20,0,2)
g2 <- rnorm(20,0,2)
g3 <- rnorm(20,0,2)
x <- c(g1,g2,g3)
group<-as.factor(c(rep(1,20),rep(2,20),rep(3,20)))
library(Rcmdr)
2012 Apr 14
2
how to divide data by week
Dear list users,
I have a data frame as below specified.
From the 1st of May to the 30th of September of several years (e.g. from 2004 to 2011) I have a frequency of accidents.
I need the mean of accidents divided by weeks (i.e. the mean of accidents from the 1st to the 7th of May of all the years, from the 8th to the 14th of May,
..., from the 29th to the 31st of May, from the 1st to the 7th of
2013 Nov 07
2
[Bug 1285] provide fallback options /etc/ssh/ssh_config
https://bugzilla.mindrot.org/show_bug.cgi?id=1285
Vladimir Rutsky <rutsky.vladimir at gmail.com> changed:
What |Removed |Added
----------------------------------------------------------------------------
CC| |rutsky.vladimir at gmail.com
--
You are receiving this mail because:
You are watching the assignee of
2011 Oct 13
5
Counting the number of integers at one swoop
Dear R users,
I'd like to count the number of integers in a vector y.
Here is an example.
y <- c(0,1,1,3,3,3,5,5,6)
In fact, I know how to count the number of specific number in y.
sum(y==0) -> 1
sum(y==1) -> 2
sum(y==2) -> 0
sum(y==3) -> 3
sum(y==4) -> 0
sum(y==5) -> 2
sum(y==6) -> 1
However, in one computation I want to get this vector [1,2,0,3,0,2,1].
Thank
2012 Jun 25
4
do.call or something instead of for
Dear R users,
I'd like to compute X like below.
X_{i,t} = 0.1*t + 2*X_{i,t-1} + W_{i,t}
where W_{i,t} are from Uniform(0,2) and X_{i,0} = 1+5*W_{i,0}
Of course, I can do this with "for" statement, but I don't think it's good
idea because "i" and "t" are too big.
So, my question is that
Is there any better idea to avoid "for" statement
2019 Aug 06
2
Dovecot Sieve doesn't move spam to Junk folder in maildir
I have an *Ubuntu 16.04* mail server with
postfix/spamassassin/*dovecot-2.2.2**2* with virtual mailboxes in maildir
format. The whole process is working excellently since a year. Spams are marked
with "*****SPAM*****" in subject and with "X-Spam-Flag: YES" in email header.
I tried now to add sieve to dovecot to move the SPAM messages into the Junk
folder without any
2013 Jun 27
3
using "rollapply" to calculate a moving sum or running sum?
#using "rollapply" to calculate a moving sum or running sum?
#I am tryign to use rollapply to calcualte a moving sum? #I tried
rollapply and get the error message
#"Error in seq.default(start.at, NROW(data), by = by) :
# wrong sign in 'by' argument"
#example:
mymatrix <- ( matrix(data=1:100, nrow=5, ncol=20) )
mymatrix_cumsum <- ( matrix(data=NA, nrow=5,
2009 Mar 23
1
performance: zoo's rollapply() vs inline
zoo's rollapply() function appears to be extremely useful for plugging in a
function on-the-fly to run over a window. With inline, there is a lot more
coding and room for error, and the code is less portable because the user
has to have R compiling set up or it won't work.
However, rollapply() seems to be really slow. Several orders of magnitude
slower than inline, in fact. I don't
2002 Dec 09
1
ifelse ?
Hi,
i want transform data and using apply(data,2,fuz)
, but i don't know why the values < low & > high don't get 0 or 1, they
get
the value from Formula ,too "((x-low)/(high-low))" what's not the
intention ?
....is switch more approriate !?
thanks for advance ,Christian
fuz <- function (x) {
#x <- na.omit(x)
low <- quantile(x,0.15)[[1]]
high
2010 Apr 09
3
"fill in" values between rollapply
Hi,
Sorry ahead of time for not including data with this question.
Using rollapply to calculate mean values for 5 day blocks, I'd use this:
Roll5mean <- rollapply(data, 5, mean, by=5, align = c("left"))
My question is, can someone tell me how to fill in the days between each of
these means with the previously calculated mean? If this doesn't make
sense, I will clarify and