search for: colun

Displaying 9 results from an estimated 9 matches for "colun".

Did you mean: colon
2008 Aug 13
1
Help me: nls and try function
Dear All, I have these problems: 1) How can use the function try in nls model: try(nls(...)) 2) I have 100 colun with data and I want ro prepare 99 file with the first colun with the others Time A1 A2 A3 A4 AN..... I want to have 99 files with a)Time and A1 b)Time and A2 n) Time AN thanks for any help M
2009 Aug 06
2
problem with r import data
...n use L<- read.csv2("path of file") now i see that the dimension of my table importing into R is 1 colonne (i have realy 20) and (655555) the number of all my lines in the excel file (i want to have only the 2600 lines in my table) Also, when i say to do this L[c(1,2,5)] it is not work coluns not define!!!! someone have an idea? thanks [[alternative HTML version deleted]]
2018 May 10
2
using for loop with data frames.
...ta frames, all with same variables df_bs_id1 <- read.csv("test1.csv",header =TRUE) df_bs_id2 <- read.csv("test2.csv",header =TRUE) df_bs_id3 <- read.csv("test3.csv",header =TRUE) How could I would implement a code loop that , for instance, would select two coluns of interest in a fashion of the code below ? # selecting only 2 columns of interest for (1, 1:3) { df_selected [i] <- df_bs_id[i] [ , c("column1", "column2")] } Tks MMS [[alternative HTML version deleted]]
2004 Apr 02
3
Doing SQL GROUP BY in R
I want a list of the number of times some factor levels appear together, similar to the following SQL statement: SELECT A, B, COUNT(C) FROM TBL GROUP BY A, B How do I do that with a data.frame in R? Thanks, Jesper Frickmann Statistician, Quality Control Novozymes North America Inc. 77 Perry Chapel Church Road Franklinton, NC 27525 USA Tel. +1 919 494 3266 [[alternative HTML version deleted]]
2008 Jul 01
3
Ignore blank columns when reading a row from a table
Hi, I am extracting data from a table where the rows have different column lengths, and empty columns have NA in them. Whenever I extract a row with some empty columns, the resulting vector carries all the NAs. Is there a way to ignore the empty columns? Thanks, -Nina
2007 Mar 30
1
faster computation of cumulative multinomial distribution
Dear list members, I have a series of /unequal/ probabilities [p1,p2,...,pk], describing mutually exclusive events, and a "remainder" class with a probability p0=1-p1-p2-....-pk, and need to calculate, for a given number of trials t>=k, the combined probability that each of the classes 1...k contains at least 1 "event" (the remainder class may be empty). To me this reaks
2002 Nov 07
1
smbpasswd (adding machines?) with pdb_mysql doesn't work
...gontime mysql:logoff time column = logofftime mysql:kickoff time column = kickofftime mysql:username column = username mysql:nt username column = ntusername; mysql:fullname column = fullname; mysql:workstations column = workstations; mysql:uid column = uid; mysql:gid colun = gid; mysql:user sid column = usersid; mysql:group sid column = groupsid; mysql:lanman pass column = lanmanpass; mysql:nt pass column = ntpass; mysql:plaintext pass column = plaintextpass; mysql:pass last set time column = passlastset; mysql:pass can change time column...
2002 Nov 07
1
pdb_mysql isn't working for me :-(
...gontime mysql:logoff time column = logofftime mysql:kickoff time column = kickofftime mysql:username column = username mysql:nt username column = ntusername; mysql:fullname column = fullname; mysql:workstations column = workstations; mysql:uid column = uid; mysql:gid colun = gid; mysql:user sid column = usersid; mysql:group sid column = groupsid; mysql:lanman pass column = lanmanpass; mysql:nt pass column = ntpass; mysql:plaintext pass column = plaintextpass;
2012 Sep 27
4
Colsplit, removing parts of a string
Hi, I am using colsplit (package = reshape) to split all strings in a column according to the same patterns. Here an example: library(reshape2) df1 <- data.frame(x=c("str1_name2", "str3_name5")) df2 <- data.frame(df1, colsplit(df1$x, pattern = "_", names=c("str","name"))) This is nearly what I want but I want to remove the words