Displaying 20 results from an estimated 10000 matches similar to: "Assigning factor to character vector"
2013 May 21
1
Lattice, ggplot, and pointsize
Hi!
When inserting R plots into a document using odfWeave, I fought for a
while to get Lattice plots use the same text size as base plots. I
eventually discovered that specifying a point size via e.g.
svg(pointsize=10) has no effect on Lattice plots. One needs to adjust
the size manually via:
trellis.par.set(fontsize=list(text=10, points=8))
This is also developed for both Lattice and ggplot2 by
2013 Feb 21
4
Getting htmlParse to work with Hebrew? (on windows)
Hello dear R-help mailing list.
Looks like the same issue in Russian:
library(RCurl)
library(XML)
u = " http://www.cian.ru/cat.php?deal_type=2&obl_id=1&room1=1"
a = getURL(u)
a # Here - the Russian is fine.
a2 <- htmlParse(a)
a2 # Here it is a mess...
None of these seem to fix it:
htmlParse(a, encoding = "windows-1251")
htmlParse(a, encoding =
2015 Jun 16
4
Bugzilla activity?
Hi!
I was wondering whether anybody was looking at the bugs on Bugzilla. I'm
asking because I've seen bugs tackled on the mailing list quite quickly,
but two fully reproducible reports I've filed on Bugzilla haven't
triggered any reaction in several weeks (for the older one).
FWIW, these are:
- Line goes beyond plot region
2012 Sep 10
1
E-mail # 2 / attachments matrix test cases
HI Sridhar,
Try this:
#I saved the file as .csv.
?testMaster<-read.csv("test1Master.csv")
?test2Sorted<-read.csv("test2Sorted.csv")
testMaster1<-testMaster[order(testMaster[,1]),]
test2Sorted1<-test2Sorted[order(test2Sorted[,1]),]
Combinedat<-as.matrix(merge(testMaster1,test2Sorted1,by="X"))
?rownames(Combinedat)<-Combinedat[,1]
2013 Apr 11
3
odfWeave: Some questions about potential formatting options
Hello All,
Learning to use the odfWeave package. I really like the package. It has good documentation, makes some very nice looking tables, and seems to have lots of options for customizing output.
There are a few things I'd like to do that don't seem to be covered in the documentation though. So I'm not sure if they're possible or not.
Here's a list of some things I'd
2017 Jan 19
2
xtabs(), factors and NAs
Hi all,
I know this issue has been discussed a few times in the past already,
but Martin Maechler suggested in a bug report [1] that I raise it here.
Basically, there is no (easy) way of printing NAs for all variables
when calling xtabs() on factors. Passing 'exclude=NULL,
na.action=na.pass' works for character vectors, but not for factors.
> test <-
2012 Jan 13
3
tm package, custom reader
I need help with creating custom xml reader for use with the tm package. The
objective is to crate a corpus for analysis. Files that I'm working with
come from solr and are in a funky XML format never the less I'm able to
parse the XML files using solrDocs.R function provided by Duncan Temple
Lang.
The problem I'm having that once I parse the document I need to create a
custom
2013 Sep 30
4
read.table() with quoted integers
Hi!
It seems that read.table() in R 3.0.1 (Linux 64-bit) does not consider
quoted integers as an acceptable value for columns for which
colClasses="integer". But when colClasses is omitted, these columns are
read as integer anyway.
For example, let's consider a file named file.dat, containing:
"1"
"2"
> read.table("file.dat",
2011 Dec 20
2
Extract BIC for coxph
Dear all,
is there a function similar to extractAIC based on which I can extract the
BIC (Bayesian Information Criterion) of a coxph model?
I found some functions that provide BIC in other packages, but none of them
seems to work with coxph.
Thanks,
Michael
[[alternative HTML version deleted]]
2012 Oct 22
1
What is behind class coercion of a factor into a character
Hello all,
Please review the following simple code:
# make a factor:
x <- factor(c("one", "two"))
# what should be the output to the following expression?
c(x, "3") # <=== ????
# I expected it to be as the output of:
c(as.character(x), "3")
# But in fact, the output is what would happen if we had ran the
next line:
2012 Oct 25
2
Regarding the memory allocation problem
Dear All,
My main objective was to compute the distance of 100000 vectors from a
set having 900 other vectors. I've a file named "seq_vec" containing
100000 records and 256 columns.
While computing, the memory was not sufficient and resulted in error
"cannot allocate vector of size 152.1Mb"
So I've approached the problem in the following:
Rather than reading the data
2013 May 26
1
load ff object in a different computer
Hi all,
I am having trouble loading a ff object previously saved in a different
computer. I have both files .ffData and .RData, and the first of them is
13Mb large from which I know the data is therein. But when I try to ffload
it,
checkdir error: cannot create /home/_myUser_
Permission denied
unable to process
home/_myUser_/Rtempdir/ff1a831d500b8d.ff.
and
2012 Oct 27
1
contr.sum() and contrast names
Hi!
I would like to suggest to make it possible, in one way or another, to
get meaningful contrast names when using contr.sum(). Currently, when
using contr.treatment(), one gets factor levels as contrast names; but
when using contr.sum(), contrasts are merely numbered, which is not
practical and can lead to mistakes (see code at the end of this
message).
This issue was discussed quickly in 2005
2014 Dec 15
3
Making iconv portable?
Hello, All:
What would it take to make ?iconv? portable?
I ask, because I want to convert accented characters to vanilla ASCII, thereby converting, e.g., ?Ra?l? to ?Raul?, and Milan Bouchet-Valet suggested on R-help that I use 'iconv(x, ?", "ASCII//TRANSLIT?)?. This worked under Windows but failed on Linux and Mac. It?s part of the ?subNonStandardCharacters? function
2011 Nov 16
1
"Non-finite finite-difference value" error in eha's aftreg
Hi list!
I'm getting an error message when trying to fit an accelerated failure
time parametric model using the aftreg() function from package eha:
> Error in optim(beta, Fmin, method = "BFGS", control = list(trace =
> as.integer(printlevel)), :
> non-finite finite-difference value [2]
This only happens when adding four specific covariates at the same time
in the
2012 Feb 10
1
Out of date instructions to build R using MKL
Hi!
I've been playing with MKL for a few days and I noticed the instructions
in the R Installation Administration manual [1] no longer apply. It
seems that since version 10.0 (the one used by the manual),
libmkl_lapack.so has been renamed/split (although the official
explanations seem to imply this was already the case in 10.0 [2]).
As a consequence, the instructions for dynamic linking no
2012 May 08
2
Dividing tick-data into intervalls
Hi everybody, I am sorry that I am kind of spamming this forum, but I have searched for some input everywhere and cant really find a nice solution for my problem.
Data looks like:
price
2011-11-01 08:00:00 0.000000000
2011-11-01 08:00:00 0.000000000
2011-11-01 08:02:00 0.000000000
2011-11-01 08:03:00 -0.017033339
2011-11-01 08:13:00 0.000690001
2011-11-01
2014 Dec 15
2
Making iconv portable?
> On Dec 15, 2014, at 12:21 PM, Kurt Hornik <Kurt.Hornik at wu.ac.at> wrote:
>
>>>>>> Spencer Graves writes:
>
>> Hello, All:
>> What would it take to make ?iconv? portable?
>
>
>> I ask, because I want to convert accented characters to
>> vanilla ASCII, thereby converting, e.g., ?Ra?l? to ?Raul?, and
>> Milan
2015 Apr 29
2
Formula evaluation, environments and attached packages
Hi!
Some time ago, I replaced calls to library() with calls to
requireNamespace() in my package logmult, in order to follow the new
CRAN policies. But I just noticed it broke jackknife/bootstrap using
several workers via package parallel.
The reason is that I'm running model replicates on the workers, and the
formula includes non-standard terms like Mult() which are provided by
gnm. If gnm
2013 Mar 18
4
!0 + !0 == !0 - !0
Hi all,
The subject line is TRUE.
Today I accidentally typed rnorm(!0).
My old eyes took a minute to focus clearly enough to see what I really typed and
why I got '!0' random numbers instead of '10' random normal numbers.
If the subject line is disturbing, be assured that this is TRUE:
!0^2 == !0 * !0 # ;-)
Anyway, I hope the hands who have been around long enough to know