similar to: Using a loop to define new variables

Displaying 20 results from an estimated 90 matches similar to: "Using a loop to define new variables"

2010 Oct 06
3
Adding a time variable prior to using reshape
Hi, I want to use reshape to convert from a skinny to a wide data format. My data doesn't have a time variable attached - I have a series of ordered observations for each subject, and it is this ordering that I am interested in (my objective is to model the most recent observation based on the preceding observations). From my understanding, prior to using reshape I have to attach a time
2011 Nov 04
2
Select some, but not all, variables stepwise
Hi, I would like to fit a linear model where some but not all explanators are chosen stepwise - ie I definitely want to include some terms, but others only if they are deemed significant (by AIC or whatever other approach is available). For example if I wanted to definitely include x1 and x2, but only include z1 and z2 if they are significant, something like this: df <-
2008 Aug 27
1
ifelse() fill order and recycling rules [Sec=Unclassified]
Hi all, Using R v2.7.1, platform i386-pc-mingw32 Can someone please shed some light on the behaviour of ifelse() for me? My intent is to calc relative proportions of z$b, at the same time subsetting z$b based on z$a. I could attack the problem other ways (suggestions welcome) but I am also intrigued by the _order_ in which ifelse seems to assign values, and how recycling works. For instance,
2011 Oct 24
4
Lm function: Error in model.frame.default
Hello, I am trying to get a linear model of y ~ log(x). *> lm (y~log(x))* However, I always get an error report: /Error in model.frame.default(formula = y ~ log(x), drop.unused.levels = TRUE) : variable lengths differ (found for 'log(x)')/ *Here was my y:* > y [1] 0.4500000 0.0500000 0.5000000 0.4000000 0.0000000 0.5000000 0.4000000 [8] 0.0500000
2009 Jun 07
2
Need some help in R : value more than equals to a row.
Hallo, I was trying some code, but couldn't make one step of the code properly. Can anybody please help me? I have one matrix like this > values [,1] [,2] [,3] [,4] [,5] [1,] 0.7777778 0.36111111 0.22222222 0.1388889 0.0000000 [2,] 1.0000000 0.00000000 0.53846154 0.0000000 0.5384615 [3,] 0.5200000 0.48000000 0.64000000 0.0000000 0.8800000 [4,] 0.8928571
2009 Jul 26
2
problems hist() and density
Hello, I have a problem with the hist() function and showing densities. The densities sum to 50 and not to 1! I use R version 2.9.1 (2009-06-26) and I load the seqinR library. My data is the following vector: [1] 0.1400000 0.2000000 0.2200000 0.2828283 0.1600000 0.1600000 0.3600000 [8] 0.1600000 0.2200000 0.2600000 0.2000000 0.3000000 0.2200000 0.2342342 [15] 0.1800000 0.2200000 0.1600000
2006 Feb 08
1
adding variable into dataframe by indice
R-Helpers, I am trying to insert a value into a dataframe. This value is a proportion calculated by counting the number of those individuals with that value and then inserting the proportion at the end of the dataframe to only those individuals with the given value. The problem I am running into is that the proportions are not being attached to only those individuals with the specified value
2009 Jun 10
1
R: Best way to plot a Matrix of all possible pair combinations
Hallo R Users, Please help I have some distance matrix data like > M[1:10,] [,1] [,2] [,3] [,4] [,5] [,6] [1,] 0.8750000 0.5000000 0.5000000 0.3750000 0.6250000 0.00000000 [2,] 0.8928571 1.0000000 0.0000000 0.8928571 0.1071429 0.00000000 [3,] 0.8928571 1.0000000 0.0000000 0.8928571 0.1071429 0.00000000 [4,] 0.8928571 1.0000000 0.0000000 0.8928571
2010 Feb 15
4
density estimates for fixed points
Problem: Based on a n x 2 data matrix i want a kernel estimate of the bivariate density. However, i also wish to specify wich points the density should be calculated at. I can offcourse just write the full kernel density estimate as a R-code, but surely there must already exist some package for this operation? The package density(), seems to create a new matrix (depending on n), where the
2007 May 05
1
(no subject)
Dear Mailing-List, I think this is a newbie question. However, i would like to integrate a loop in the function below. So that the script calculates for each variable within the dataframe df1 the connecting data in df2. Actually it takes only the first row. I hope that's clear. My goal is to apply the function for each data in df1. Many thanks in advance. An example is as follows: df1
2007 Mar 28
2
Standardization Range
? stato filtrato un testo allegato il cui set di caratteri non era indicato... Nome: non disponibile Url: https://stat.ethz.ch/pipermail/r-help/attachments/20070328/ee4422a2/attachment.pl
2024 Apr 13
1
Just for your (a|be)musement.
I recently had the chance to read a book explaining how to use ChatGPT with a certain programming language. (I'm not going to describe the book any more than that because I don't want to embarrass whoever wrote it.) They have appendix material showing three queries to ChatGPT and the answers. Paraphrased, the queries are "if I throw 2 (3, 4) fair dice, what is the probability I get
2016 Apr 24
3
Dividing rows in groups
Hi I have two data frames as shown below (second one is obtained by aggregating rows of similar IDs in df1.). They both have similar number of columns but rows of df2 are lesser than rows of df1. df1: ID A B 1 1 2 1 0 3 2 5 NA 2 1 3 3 1 4 4 NA NA 4
2011 Oct 05
1
Dealing with proportions
Dear list, I have very little experience in dealing with proportions, i am sure this is a very simple question but i could find no suitable answer beyond doing a chi-sq test and then using the Marascuilo procedure as a post-hoc analysis. I am simply wanting to know if the proportions ( i.e the number of Yes / No) significantly differ between the cases and if so which cases are significantly
2002 Mar 13
0
rpart error with 0-frequency factor levels (with partial fix) (PR#1378)
(I'm sending to r-bugs because rpart is one of the recommended packages and is always installed. I'm also sending it directly to Dr. Ripley, as the maintainer.) rpart working as a classifier does not work (produces no splits) when the class indicator has no instances of one of the factor levels, as long as the factor level is not the final level. I have at least a partial fix, which I
2012 Oct 25
5
trying ti use a function in aggregate
Hi -I am using R v 2.13.0. I am trying to use the aggregate function to calculate the percent at length for each Trip_id and CommonName. Here is a small subset of the data. Trip_id Vessel CommonName Length Count 1 230 Sunlight Shad,American 19 1 2 230 Sunlight Shad,American 20 1 3 230 Sunlight Shad,American 21
2016 Apr 24
0
Dividing rows in groups
This will handle all the columns; it assumes the ones you want to start with are in column 2 through the end: > library(dplyr) > df1 <- read.table(text = "ID A B + 1 1 2 + 1 0 3 + 2 5 NA + 2 1 3 + 3 1 4 + 4 NA NA + 4 0 1 + 4
2006 Apr 17
0
Problem getting R's decision tree for Quinlan's golf exam ple data [Broadcast]
See ?rpart.control. I get: > golf.rp = rpart(Outlook ~ ., golf, control=rpart.control(minsplit=1)) > golf.rp n= 14 node), split, n, loss, yval, (yprob) * denotes terminal node 1) root 14 9 rain (0.2857143 0.3571429 0.3571429) 2) Temperature< 71.5 6 2 rain (0.1666667 0.6666667 0.1666667) 4) Temperature< 64.5 1 0 overcast (1.0000000 0.0000000 0.0000000) * 5)
2004 Aug 10
0
Check failed after compilation (PR#7159)
Full_Name: Madeleine Yeh Version: 1.9.1 OS: AIX 5.2 Submission from: (NULL) (151.121.225.1) After compiling R-1.9.1 on AIX 5.2 using the IBM cc compiler, I ran the checks. One of them failed. Here is the output from running the check solo. root@svweb:/fsapps/test/build/R/1.9.1/R-1.9.1/tests/Examples: ># ../../bin/R --vanilla < stats-Ex.R R : Copyright 2004, The R
2013 Feb 15
10
reading data
Hi, #working directory data1 #changed name data to data1.? Added some files in each of sub directories a1, a2, etc. ?indx1<- indx[indx!=""] lapply(indx1,function(x) list.files(x)) #[[1]] #[1] "a1.txt"??????? "mmmmm11kk.txt" #[[2]] #[1] "a2.txt"??????? "mmmmm11kk.txt" #[[3]] #[1] "a3.txt"??????? "mmmmm11kk.txt" #[[4]] #[1]