Displaying 20 results from an estimated 22 matches for "fnams".
Did you mean:
fname
2001 Nov 01
3
Lost backslashes in parse()
I'm comparing R-1.3.0 on Solaris 2.6 to R-1.3.1 on WindowsNT. The following
5-line script returns TRUE on Unix but FALSE on NT:
fnam <- tempfile()
file.create(fnam)
txt <- paste("file.exists(\"", fnam, "\")", sep="")
expr <- parse(text=txt)
eval(expr)
The issue seems to be that backslashes get lost in the parsing. A workaround
is to change
2007 Mar 09
1
Applying some equations over all unique combinations of 4 variables
...lculations need to
be done on one factor at a
#I then have a set of different rates that are applied
#to it.
#dataset
cata <- c( 1,1,6,1,1,2)
catb <- c( 1,2,3,4,5,6)
doga <- c(3,5,3,6,4, 0)
data1 <- data.frame(cata, catb, doga)
rm(cata,catb,doga)
data1
# start rates
# names for lists
fnams <- c("af", "pf", "cf", "mf")
mnams <- c("am", "pm", "cm", "mm")
# Current layout of the rate data frames
alphahill <- list(af <- c("a1","a2","a3"), pf <-
c("d1"...
2013 Apr 10
1
Issue with Control-Z in a text file on Windows - readLines() appears to truncate
Working on Windows I have had to deal with CSV files that,
unfortunately, contain embedded Control-Zs, i.e. ASCII character 26 in
decimal, and the readLines() function in R on Windows (2.15.2 and
3.0.0) appears to truncate at the control-Z. There is no problem at
all on Ubuntu Linux with R 3.0.0.
Am I mistaken or is this genuine?
# Create a small file with embedded Control-Z
h3 <-
2012 Sep 26
3
Reading multiple files
Hi,
I have 35 data files for reading. I would like get a program for
performing reading of 35 files at once.
All are of the type: Dados1.raw, Dados2.raw and so on.
If the files have the same number of columns, I can read with the
following commands:
rm(list=ls())
filenames = list.files(path="~/Silvano/Arq", pattern="Dados+.*raw")
names = substr(filenames, 1, 7)
for(i in
2010 May 07
0
A fix that for 'bquote' that may work (PR#14031)
--- On Fri, 6/11/09, tlumley at u.washington.edu <tlumley at u.washington.edu> wrote:
> From: tlumley at u.washington.edu <tlumley at u.washington.edu>
> Subject: Re: [Rd] A fix that for 'bquote' that may work? (PR#14031)
> To: suharto_anggono at yahoo.com
> Cc: r-devel at stat.math.ethz.ch, R-bugs at r-project.org
> Date: Friday, 6 November, 2009, 11:42 PM
>
2011 Feb 14
3
R command line and pipe using in Linux?
Hi,
I have a very large data file(GB) from which I only want to extract one
column to draw histogram. This would be done several times, so I would like
to ask if there is anyway to plot this using R from the linux command line,
something look like this
cut -f1 xxx.txt |RplotHist ....
Thanks and hope to hear from you.
Best regards,
Hang
[[alternative HTML version deleted]]
2012 Jan 17
2
How to loop on file names
Dear all,
I need to do the same procedure on several files. But I don't know how
to refer to the file name.
Here is an example of what I am trying to do.
List of files: file1(A,B,C, D1...Dn), file2(A,B,C,E1,...,En),
file3(A,B,C,F1,...,Fn)
Procedure I want to apply on each file:
dft <- melt(df,id=c('A','B','C'))
dft$X <- substr(dft$variable,1,3)
dft$Y <-
2019 Mar 11
0
R 3.5.3 is released
The build system rolled up R-3.5.3.tar.gz (codename "Great Truth") this morning.
The list below details the changes in this release. This is the wrap-up release for the 3.5.x series, so actually, not much has happened.
You can get the source code from
http://cran.r-project.org/src/base/R-3/R-3.5.3.tar.gz
or wait for it to be mirrored at a CRAN site nearer to you.
Binaries for
2019 Mar 11
0
R 3.5.3 is released
The build system rolled up R-3.5.3.tar.gz (codename "Great Truth") this morning.
The list below details the changes in this release. This is the wrap-up release for the 3.5.x series, so actually, not much has happened.
You can get the source code from
http://cran.r-project.org/src/base/R-3/R-3.5.3.tar.gz
or wait for it to be mirrored at a CRAN site nearer to you.
Binaries for
2019 Mar 11
0
R 3.5.3 is released
The build system rolled up R-3.5.3.tar.gz (codename "Great Truth") this morning.
The list below details the changes in this release. This is the wrap-up release for the 3.5.x series, so actually, not much has happened.
You can get the source code from
http://cran.r-project.org/src/base/R-3/R-3.5.3.tar.gz
or wait for it to be mirrored at a CRAN site nearer to you.
Binaries for
2013 Apr 09
1
Fwd: R unzip method gives filenames as character
Question: would it be better if the contents list of the unzip() function
returned the filenames as character rather than factor since they are
probably unique strings?
> tmp <- unzip(fnam, list=TRUE)
> str(tmp)
'data.frame': 31 obs. of 3 variables:
$ Name : Factor w/ 31 levels "fred1.csv",..: 1 2 3 4 5 6 7 8 9 10 ...
$ Length: num 424486 2664277 219798 442383
2011 Aug 16
1
PBSmapping, where is Ireland?!
Hi folks,
I've been using 'PBSmapping' to make a map of Europe with some labels. I've
been using the 'worldLL' PolyData, as my computer is too slow to make my own
from the GSHHS files.
The only problem is this PolyData does not seem to include Ireland. I have
no idea why this should be so, other European islands such as Sardinia etc.
are included.
2010 Jan 09
4
parsing pdf files
I have a pdf file that I would like to parse into R:
http://www.williams.edu/Registrar/geninfo/faculty.pdf
For now, I open the file in Acrobat by hand, then save it "as text"
and then use readLines(). That works fine but a) I am concerned that
some information may be lost and b) I may be doing this a lot, so I
would rather have R grab the information from the pdf file directly.
So: is
2009 Sep 10
2
"Read.csv" in R with dynamic file (1st) argument
Dear R users,
I have numerous data sets (csv files) saved in the folder which has the same
name as individual data.
(i.e data x1 saved in x1 folder, data x2 in x2 folder etc)
I would like to read in the desired data set name using 'scan' function and
assign this inputted value to an object so that it can be used in the
'read.csv' function.
For example,
x <- scan()
1: 0708
2009 Nov 09
3
Hand-crafting an .RData file
Hello,
I frequently have to export a large quantity of data from some
source (for example, a database, or a hand-written perl script) and then
read it into R. This occasionally takes a lot of time; I'm usually using
read.table("filename",comment.char="",quote="") to read the data once it is
written to disk.
However, I *know* that the program that generates
2004 Jan 27
1
Differentiating debug messages from both sides
Some of the debug messages that rsync outputs (when verbose >= 2) can
occur on both sides of the connection. This makes it hard to know which
program is saying what. Some debug messages deal with this by
outputting a "[PID]" string at the start of the message. Unfortunately,
the startup message that tells us which pid is which is only output when
verbose >= 3, so there's a
2004 Jan 19
1
File that "vanish"es between readdir and stat is not IO error
Using rsync 2.6.0 with --verbose and doing a pull.
>?receiving file list ... readlink "{FILENAME}" failed:
>?No such file or directory
>?done
>?IO error encountered - skipping file deletion
The file was a temporary file that was being deleted just as
the rsync was run. So while the file list was being built,
it was there when the directory was read but had vanished
by the
2004 May 02
1
SEGV on FreeBSD 4.8-STABLE with 2.6.2
I'm getting a SEGV on a FreeBSD 4.8-STABLE box. The client is Solaris
9/SPARC. Both boxes run 2.6.2.
The command I'm running is:
$ rsync -arHRv --numeric-ids --delete --exclude=/opt/dist/cdrom \
[paths] [server]:[path]
If I whittle down what appears in [paths], then it works.
$ gdb rsync rsync.core
gdb> bt
#0 0x280faf0d in strncmp () from /usr/lib/libc.so.4
#1 0x7 in ?? ()
#2
2004 Feb 02
1
[PATCH] --one-file-system and automounter
We use rsync in a Linux installation script. First, the root filesystem
of another machine on the network is cloned with "rsync -axzH", and then
a few files are updated to give the clone its own identity. This works
fine, but last week, the Postfix mailer daemon on a new machine refused
to start because some lock files had a link count of 2. It turned out
that rsync had created two
2004 Feb 06
4
memory reduction
As those of you who watch CVS will be aware Wayne has been
making progress in reducing memory requirements of rsync.
Much of what he has done has been the product of discussions
between he and myself that started a month ago with John Van
Essen.
Most recently Wayne has changed how the file_struct and its
associated data are allocated, eliminating the string areas.
Most of these changes have been