similar to: Removing Rows

Displaying 20 results from an estimated 30000 matches similar to: "Removing Rows"

2007 Mar 02
3
Reformulated matrices dimensions limitation problem
First I wanted to thank both Marc Schwartz Greg Snow and for their reply. Then I needed to add a level of complexity to the problem. I would be able to create the biggest possible matrix. In other way does it exist a method to ask smthing like the following : max number of rows for a matrix if column=x? Thank you ------------------------------------------------------ Passa a Infostrada.
2006 Feb 07
1
(second round) creating a certain type of matrix
Hi R users Here is what I got with help from Petr Pikal (Thanks Petr Pikal). I modified Petr Pikal's code to a little to meet my purpose. I created a function to generate a matrix generate.matrix<-function(n.variable) { mat<-matrix(0,n.variable,(n.variable/2)/5+1) #matrix of zeroes dd<-dim(mat) # actual dimensions mat[1:(dd[1]/2),1]<-1 #put 1 in first half of first column
2005 Mar 13
4
Output a dataframe from R to excel
Hi, I am trying to output an dataframe from R to Excel file. Can anyone tell me how to do it? Thanks a lot. Eg. R dataframe: A B C 1 2 1 3 4 2 . . . [[alternative HTML version deleted]]
2007 Jan 05
4
Fast Removing Duplicates from Every Column
Hi, I'm looking for some lines of code that does the following: I have a dataframe with 160 Columns and a number of rows (max 30): Col1 Col2 Col3 ... Col 159 Col 160 Row 1 0 0 LD ... 0 VD Row 2 HD 0 0 0 MD Row 3 0 HD HD 0 LD Row 4 LD HD HD 0 LD ... ... LastRow HD HD LD 0 MD Now I want a dataframe that looks like this. As you see
2017 Nov 09
2
weighted average grouped by variables
Hi Thanks for working example. you could use split/ lapply approach, however it is probably not much better than dplyr method. sapply(split(mydf, mydf$type), function(speed, n_vehicles) sum(mydf$speed*mydf$n_vehicles)/sum(mydf$n_vehicles)) gives you averages aggregate(mydf$n_vehicles, list(mydf$type), sum)$x gives you sums Cheers Petr > -----Original Message----- > From: R-help
2006 Apr 14
2
another very simple loop question
I have a dataset with 4 years of students, and normally I want to estimate things using each individual year, so I have a for loop as follows for (i in 1:4){} However, the only way I know how to calculate estimates using all four years of data is to put the estimations outside of the loop. Is there anyway to make a for loop that uses all four years at once, then uses each individual year?
2017 Nov 11
0
weighted average grouped by variables
> On 9 Nov 2017, at 14:58, PIKAL Petr <petr.pikal at precheza.cz> wrote: > > Hi > > Thanks for working example. > > you could use split/ lapply approach, however it is probably not much better than dplyr method. > > sapply(split(mydf, mydf$type), function(speed, n_vehicles) sum(mydf$speed*mydf$n_vehicles)/sum(mydf$n_vehicles)) > gives you averages > The
2005 Oct 25
1
selecting every nth item in the data
I want to make a glm and then use predict. I have a fairly small sample (4000 cases) and I want to train on 90% and test on 10% but I want to do it in slices so I test on every 10th case and train on the others. Is there some simple way to get these elements? Stephen -- 21/10/2005 [[alternative HTML version deleted]]
2004 Jun 07
3
Aggregate rows to see the number of occurences
Hi, I have a set of data like the following: [,1] [,2] [1,] 10 2 [2,] 7 0 [3,] 1 0 [4,] 1 0 [5,] 15 0 [6,] 17 4 [7,] 4 0 [8,] 19 8 [9,] 10 2 [10,] 19 5 I'd like to aggregate it in order to obtain the frequency (the number of occurences) for each couple of values (e.g.: (10,2) appears twice, (7,0) appears once). Something cool
2005 Oct 11
2
Problems with plot function
Hello all R users, My simulation function works correctly, but I have problems with plot function. You will find the following code using it. Thank you for your help ##################################################" simulation <- function(k, n){ conc <- seq(0,10,by=0.5) #choixg <- seq(1, length(conc)) choixg <- rep(0,length(conc)) for (i in 1:length(conc)){ choixg[i]
2007 Mar 08
2
Removing duplicated rows within a matrix, with missing data as wildcards
I'd like to remove duplicated rows within a matrix, with missing data being treated as wildcards. For example > x <- matrix((1:3), 5, 3) > x[4,2] = NA > x[3,3] = NA > x [,1] [,2] [,3] [1,] 1 3 2 [2,] 2 1 3 [3,] 3 2 NA [4,] 1 NA 2 [5,] 2 1 3 I would like to obtain [,1] [,2] [,3] [1,] 1 3 2 [2,] 2 1 3
2010 Oct 21
4
Efficient nested loops
Dear R community, I am working with huge arrays, so I spend a lot of time computing. This is my code: for (x in 1:dim(variable)[1]){ for (y in 1:dim(variable)[2]){ for (z in 1:dim(variable)[3]){ result <- max(variable[x,y,z,]) } } } Is there a more efficient procedure to do this task? Thanks in advance! [[alternative HTML version deleted]]
2007 Dec 18
4
accessing dimension names
I have a matrix y: > dimnames(y) $x93 [1] "1" "2" $x94 [1] "0" "1" "2" .................. so on (there are other dimensions as well) I need to access a particular dimension, but a random mechanism tells me which dimension it would. So, sometimes I might need to access dimnames(y)$x93, some other time it would be dimnames(y)$x94.. and so
2018 Feb 12
2
plotting the regression coefficients
Hi Petr and Richard; Thanks for your responses and supports. I just faced a different problem. I have the following R codes and work well. p <- ggplot(a, aes(x=Phenotypes, y=Metabolites, size=abs(Beta), colour=factor(sign(Beta)))) + theme(axis.text=element_text(size = 5)) p1<-p+geom_point() p2<-p1+theme(panel.grid.major = element_blank(), panel.grid.minor = element_blank(),
2012 Oct 30
6
standard error for quantile
Dear all I have a question about quantiles standard error, partly practical partly theoretical. I know that x<-rlnorm(100000, log(200), log(2)) quantile(x, c(.10,.5,.99)) computes quantiles but I would like to know if there is any function to find standard error (or any dispersion measure) of these estimated values. And here is a theoretical one. I feel that when I compute median from given
2018 Feb 12
3
plotting the regression coefficients
Hi After melt you can change levels of your factor variable. Again with the toy example. > levels(temp$variable) [1] "y1" "y2" "y3" "y4" > levels(temp$variable) <- levels(temp$variable)[c(2,4,1,3)] > levels(temp$variable) [1] "y2" "y4" "y1" "y3" > And you will get graphs with this new levels ordering.
2008 Jan 10
6
4 dimensional graphics
Dear all I want to display 4 dimensional space by some suitable way. I searched CRAN and found miscellaneous 3 dim graphics packages which I maybe can modify but anyway I am open to any hint how to efficiently display data like: longitude, latitude, height, value Thank you Petr Pikal petr.pikal at precheza.cz
2018 Feb 12
0
plotting the regression coefficients
Petr, there was a thinko in your response. tmp <- data.frame(m=factor(letters[1:4]), n=1:4) tmp tmp$m <- factor(tmp$m, levels=c("c","b","a","d")) ## right tmp[order(tmp$m),] tmp <- data.frame(m=factor(letters[1:4]), n=1:4) levels(tmp$m) <- c("c","b","a","d") ## wrong tmp[order(tmp$m),] changing levels
2018 Feb 13
0
plotting the regression coefficients
Hi scale_colour_gradient(?red?, ?blue?) should do the trick. Actually I found it by Google ggplot colour http://www.cookbook-r.com/Graphs/Colors_(ggplot2)/ http://www.sthda.com/english/wiki/ggplot2-colors-how-to-change-colors-automatically-and-manually#gradient-colors-for-scatter-plots question. So you could find it too and probably far more quickly then myself as I have also other duties. Cheers
2007 May 07
4
Mardia's multivariate normality test
Dear all, I got this error message > library(dprep) > mardia(Savg) Error in cov(data) : 'x' is empty But with the same data, I got > library(mvnormtest) > mshapiro.test(Savg) Shapiro-Wilk normality test data: Z W = 0.9411, p-value = 0.6739 What does the error message "Error in cov(data) : 'x' is empty" mean? Thanks a lot! Jiao