Displaying 20 results from an estimated 10000 matches similar to: "merge warning is.na(out$h)"
2009 Feb 26
2
Merge question
Hi:
I am a new R user. I have the following question and would appreciate your input
Data1 (data frame 1)
p1,d1,d2 (p1 is text and d1 and d2 are numeric)
xyz,10,25
Data2 (data frame 2)
p1,d1,d2
xyz,11,15
Now I want to create a new data frame that looks like so below. The fields d1 and s2 are summed by the product key.
Data3
p1,d1,d2
xyz,21 (sum of 10 from Data1 and 11 from Data2),40 (sum of 25
2003 Aug 07
2
Question about 'NA'
Hi all,
Ive got a database with 10 columns (different
variables) for 100 subjects, each column with
different # of NA's. I'd like to know if it is
possible to use a function to exclude the NA's using
only a specific column, lets say:
Data2 <- omit.exclude(Data1$column1) ??, then
Data3 <- omit.exclude(Data1$column2) and so on
I tried the code above but with no results
Thanks
2008 May 14
4
Accessing items in a list of lists
Using R 2.6.2, say I have the following list of lists, "comb":
data1 <- list(a = 1, b = 2, c = 3)
data2 <- list(a = 4, b = 5, c = 6)
data3 <- list(a = 3, b = 6, c = 9)
comb <- list(data1 = data1, data2 = data2, data3 = data3)
So that all names for the lowest level list are common. How can I most
efficiently access all of the sublist items "a" indexed by the outer
2009 Nov 12
2
redundant factor levels after subsetting a dataset
#I have a data frame with a numeric and a character variable.
x=c(1,2,3,2,0,2,-1,-2,-4)
md=c(rep("Miller",3), rep("Richard",3),rep("Smith",3))
data1=data.frame(x,md)
#I subset this data.frame in a way such that one level of the character
variable does not appear in the new dataset.
data2=data1[x>0,]
data3=subset(data1,x>0)
#However, when I check the levels
2008 Jun 04
4
merging 3 data sets at once
Hi All,
I am looking into merging 3 data sets I know how to do that by merging data1
with data2 and then merging the result with data 3. I was wondering if it
can be done all at once so I tried,
M<-merge(data1,data2,data3, by=?ID?)
It does not work!
Any ideas?
--
View this message in context: http://www.nabble.com/merging-3-data-sets-at-once-tp17658873p17658873.html
Sent from the R help
2005 Feb 16
5
scaling axes when plotting multiple data sets
1) When adding additional data sets to a plot using "plot" followed by "lines",
is there a way to automate the scaling of the axes to allow for all data sets
to fit within the plot area?
2) I attempted to solve this by setting
xlim=c(min(c(data1,data2,data3)),max(c(data1,data2,data3)))
however, there are some NAs and Infs in these data sets, and min(data1) and
max(data1) both
2002 Dec 05
1
Passing options as lists
Hi,
I apologize if this has previously been posted. I've just subscribed to the
R-help digest.
I'm writing a plotting function that uses layout() to plot several different
plots on the same device. This function uses plot(), image(), and a custom
function that uses text(). Each cell of the layout needs different par()
parameters, so what I'd like to do is pass them as lists:
2019 Mar 28
2
NFSv4: Using fsid=0 but *not* exporting the root filesystem
Hi,
I would like to use the NFSv4 ability to create a "root" filesystem with
fsid=0, so that I don't have to refer to the whole path of the exported
filesystem when I mount it. However I do *not* want this root
filesystem to be mountable by any host. Is that possible and how?
E.g
Filesystem:
/exports/data1
/exports/data2
/exports/data3
/etc/exports:
/exports
2009 Jun 03
1
Still can't find missing data - How do I get NA in xtabs with factors?
The problem here is Table doesn't seem to have a way to weigh the data.
> ToyData
Data1 Data2 Data3 Weight
101 Sam Red Banana 1.1
102 Sam Green Banana 2.1
103 Sam Blue Orange 2.1
104 Fred Red Orange 2.1
105 Fred Green Guava 2.1
106 Fred Blue Guava 2.1
107 <NA> Red Pear 50.1
108 <NA> Green Pear 50.1
109 <NA> Blue
2008 Aug 21
1
problem merging two data sets ( one with a header and one without)
I have two set of data, Data1 and Data2 . Data1 has a header and Data2 does
not. I would like to merge the two data sets after removing some columns
from data2 .
I am having a problem merging so I had to write and read final data and
specify the ?header=F? so the merge can be done by?V1?. Is there a way to
avoid this step. The problem is when I do cbind the FinalData has different
column names
2008 Jun 05
1
problems with reading the data and merge
I have a problem with reading a file data.txt that has a header
Each row has the individual ID and then some data that are letter and
numbers such as 00
If I read the file as
data<-read.table("data.txt",sep='\t', header=T)
write.table(data,file="data1.txt", sep='\t', quote=F, col.names=T,
row.names=F)
data1.txt looks different than data.txt. In
2013 Jun 10
4
Combining CSV data
Hello R community,
I am trying to combine two CSV files that look like this:
File A
Row_ID_CR, Data1, Data2, Data3
1, aa, bb, cc
2, dd, ee, ff
File B
Row_ID_N, Src_Row_ID, DataN1
1a, 1, This is comment 1
2a, 1, This is comment 2
3a,
2006 Nov 24
1
barplot help needed
hello,
I would like to create the following barplot:
I have 4 different data sets (same length + stddev for each data point)
data1
sd1
data2
sd2
data3
sd3
data4
sd4
now, I'd like to plot in the following way:
data1[1],data2[1],data3[1],data4[1] with it's sd-values side-by-side at
one x-axis label (named "position 1") and each bar in different colors.
2010 Mar 24
3
How to use the paste function to create an already used variable
Hi there,
I have the following problem
Four data frames exist:
data1
data2
data3
data4
Now I want to write a loop and temporarily store the data1, data2,
data3, data4 in a variable called data.
I tried the following...
for (i in 1:4) {
Data <- paste("data",i,sep="")
...
..
}
but it doesn't function. I think the problem is the definition of the
mode of the pasted
2013 Jan 05
3
Need help with time series data
Dear R users,
Could you share your knowledge on following problem:
Suppose we have dataframe:
ID TIME Data1 Data2 Data3
........... Data700
1. 2013-01-01 00:00:00 34 53 66
............ 55
2. 2013-01-01 00:00:01 333 4 5
............ 50
3. 2013-01-01 00:00:02 and so
2016 Feb 05
3
[PATCH] inspect: get windows drive letters for GPT disks.
This patch updates the guestfs_inspect_get_drive_mappings API call to
also return drive letters for GPT paritions. Previously this worked
only for MBR partitions. This is achieved by matching the GPT partition
GUID with the info stored in the blob from
HKLM\SYSTEM\MountedDevices\DosDevices keys. For GPT partions this blob
contains a "DMIO:ID:" prefix followed by a 16 byte binary GUID.
2004 Aug 23
2
[ Multiple drives ]
Hello,
I have 3 hdd (120 GB, 120 GB and 80 GB) mounted on
/data1 , /data2 and /data3.
All these drives must be shared via a public access
with Samba.
For the moment, I can only share the 'data1'
directory.
[public]
path = /data1
Is there a possibility to share several disks under
the same account ?
By example :
[public]
path = /data1, /data2, /data3
Then, under Windows, I'd like
2013 Mar 26
1
Increase font size in plots
Hi
I am using violin plots (type of boxplots) and I am trying to increase the font size in the plots.
It looks like that the violin plots do not work as "normal" plots as the cex parameters are ignored.
You can have a loot at the code below
require('vioplot')
data1<-rnorm(100)
data2<-rnorm(10)
data3<-rnorm(1000)
vioplot(data1,data2,data3,cex.lab=2,cex.axis=2)
2016 Feb 05
2
[PATCHv2] inspect: get windows drive letters for GPT disks.
This patch updates the guestfs_inspect_get_drive_mappings API call to
also return drive letters for GPT paritions. Previously this worked
only for MBR partitions. This is achieved by matching the GPT partition
GUID with the info stored in the blob from
HKLM\SYSTEM\MountedDevices\DosDevices keys. For GPT partions this blob
contains a "DMIO:ID:" prefix followed by a 16 byte binary GUID.
2011 Nov 18
1
couting events by subject with "black out" windows
I large datset that includes subjects(ID), Dates and events that need to be counted. Not every date includes an event, and I need to only count one event per 30days, per subject. So in essence, I need to create a 30-day "black out" period during which time an event cannot be "counted" for each subject. The reason is that a rule has been set up, whereby a subject can only be