similar to: Regular expression to find value between brackets

Displaying 20 results from an estimated 3000 matches similar to: "Regular expression to find value between brackets"

2009 Mar 06
1
Extract dates from dataframe
Hi, I have the following dataframe: ID Dates 1 16-07-01 06-10-95 2 24/01/02 06-10-95 3 16/01/02 16/08/94 12/01/91 And I would like to extract the dates, but couple the ID's to the right dates, eg: ID Dates 1 16-07-01 1 06-10-95 2 24-01-02 2 .... I have no clue about how to get started, looks something for the
2009 Oct 26
1
regular expressions
Dear list, I have the following text to parse (originating from readLines as some lines have unequal size), st = c("START text1 1 text2 2.3", "whatever intermediate text", "START text1 23.4 text2 3.1415") from which I'd like to extract the lines starting with "START", and group the subsequent fields in a data.frame in this format: text1 text2
2007 Mar 01
4
How to read in this data format?
Hi, I recieved an ascii file, containing following information: $$ Experiment Number: $$ Associated Data: FUNCTION 1 Scan 1 Retention Time 0.017 399.8112 184 399.8742 0 399.9372 152 .... Scan 2 Retention Time 0.021 399.8112 181 399.8742 1 399.9372 153 ..... I would like to import this data in R into a dataframe, where there is a column time, the first numbers as column names, and the
2010 Jul 15
3
Access web content from within R
Hi, I have to search in an online db for registered manufacturers of raw materials. Can I use R for the following: I have a list with monograph numbers eg: l <- c(198, 731,355) Now I want to make a dataframe, containing the monograph number and the information listed under COS: Certificate holder, certificate number, Status, Type Is this possible with R? kind regards Bart -- View this
2013 Jan 24
4
sorting/grouping/classification problem?
Hi, I'm a database admin for a database which manage chromatographic results of products during stability studies. I use R for the reporting of the results in MS Word through R2wd. But now I think I need your help: suppose we have the following data frame: ID rrt Mnd Result 1 0.45 0 0.10 1 0.48 0 0.30 1 1.24 0 0.50 2 0.45 3 0.20 2 0.48 3 0.60 2 1.22 3 0.40 3
2005 Dec 18
1
Fit non-lineair 3D Data
Hi, I have a problem with fitting a model: I made a dataframe with this data: a <- 1:3 b <- 1:3 c <- c(3, 2, 3, 2, 1, 2, 3, 2, 3) df <- expand.grid(a,b) df$result <- c names(df) <- c("A","B", "result") Although I can make a graph of the data: require(lattice) wireframe(result~A*B, data=df) I can't get a model to
2011 Oct 24
2
splitting a string into words preserving blanks (using regex)
I would like to split a string into words at its blanks but also to preserve all blanks. Example: c(" some words to split ") should become c(" ", "some", " ", " words", " ", "to" , " ", "split", " ") I was not able to achieve this via strsplit() . But I am not familiar with regular
2007 May 09
1
Fitting model with response and day bias
Hi, I'm trying to fit a model which has a response bias, but also a day to day bias. If I try to simulate the data, I don't get the right values with optim, and also I can't use the function to give a prediction interval. My simulated data are: DF <- as.data.frame(cbind(x=rep(1:10,2),dag=rep(1:2,each=10))) bias <- c(-0.2,0.5) DF$y <- ((DF$x-0.1) * 5)+2 +
2007 Feb 10
4
Near function?
Hi, I have an integer which is extracted from a dataframe, which is sorted by another column of the dataframe. Now I would like to remove some elements of the integer, which are near to others by their value. For example: integer: c(1,20,2,21) should be c(1,20). I tried to write a function, but for some reason, somethings won't work x <- 1:20 near <- function(x,th) { nr <-
2008 Nov 24
3
count the cumulative for each subject
I have a data set like the following: subject visit x1 1 1 0.5 1 2 1.2 1 3 0.7 2 1 0.4 2 2 0.6 2 3 1.0 ..... where x1 is the interval between the two visits. Now I want to calculate the cumulative intervals since the beinging, for example subject visit x1 cum 1 1 0.5 0.5 1 2 1.2 0.5+1.2 1 3 0.7 0.5+1.2+0.7 2 1 0.4 0.4 2 2 0.6 0.4+0.6 2 3 1.0 0.4+0.6+1.0 ..... is there an easy to generate the
2005 Nov 04
1
t test on data frame
Hi, I have constructed a dataframe as follows: Oil <- rep(c("Oil1","Oil2","Oil3"),8) Comp <- rep(rep(c("C1","C2"),c(4,4)),3) Mth <- rep(c("M1","M1","M2","M2"),6) Meas <- rep(c(1,2),12) Result <- rnorm(24,mean=100, sd=5) df <- data.frame(Oil, Comp, Mth, Meas, Result) The same compound
2006 Jun 10
1
Maximum likelihood estimation of Regression parameters
Hi, I want to use Maximum likelihood to estimate the parameters from my regression line. I have purchased the book "Applied linear statistical models" from Neter, Kutner, nachtsheim & Wasserman, and in one of the first chapters, they use maximum likelihood to estimate the parameters. Now I want to tried it for my self, but couldn't find the right function. In the book, they give
2007 Aug 23
4
Calculating diameters of cirkels in a picture.
Hi, Maybe this is more a programming questions than a specific R-project question, but maybe there is someone who can point me in the right direction. I have a picture of cirkels which I took with a digital camera. Now I want to use the diameter of the cirkels on the picture for analysis in R. I can use pixmap to import the picture, but how do I find the outside cirkels and calculate the
2006 Jun 24
2
data frame search
Hi, I want to make a data frame which contains the positions of some searched values in another data frame. Like: Dataframe 1: 1 2 3 4 1 2 3 4 2 3 4 1 2 3 4 2 4 1 2 3 2 3 4 1 Let's say I searched on "4", then Dataframe 2 should contain: x y 1 4 1 8 2 3 2 7 3 1 3 7 I have written a routine, but it seems to me that it isn't that perfect: x<- 0
2007 May 13
2
extracting text contained in brackets ("[ ... ]") from a character string?
I have a text string that contains text within two brackets. e.g. "testdata[3]" "testdata[-4]", "testdata[-4g]", I wish to "extract" the string enclosed in brackets? What is a good way to do this? e.g. fun(testdata[3]) = '3' fun(testdata[-4g]) = '-4g' --------------------------------- Moody friends. Drama queens. Your life?
2010 Mar 15
2
tcltk and R
I have had some comments on sqldf regarding its dependence on tcltk such as the second last sentence on this blog post: http://translate.google.com/translate?hl=en&sl=zh-CN&u=http://www.wentrue.net/blog/%3Fp%3D453&prev=http://blogsearch.google.com/blogsearch%3Fhl%3Den%26ie%3DUTF-8%26q%3Dsqldf%26lr%3D%26sa%3DN%26start%3D10 sqldf does not directly use tcltk but it does use strapply in
2010 Jul 02
3
Good Package(s) for String and URL processing?
Are there packages that allow improved String and URL processing? E.g. extract parts of a URLs such as sub-domains, top-level domain, protocols (e.g. https, http, ftp), file type based on endings, check if a URL is valid or not, etc... I am currently only using split and paste. Are there better and more efficient ways to handle strings e.g. finding sub-strings or to do pattern matching? What
2010 Aug 26
5
Quick GREP challenge
> grep("f[0-9]+=", "f1=5,f22=3,", value = T) [1] "f1=5,f22=3," How do I make the line output c("f1", "f22") instead? (Actually, c(1,22) would be even better). Thank you. -- View this message in context: http://r.789695.n4.nabble.com/Quick-GREP-challenge-tp2339486p2339486.html Sent from the R help mailing list archive at Nabble.com.
2008 Jan 23
1
Package Building and Name Space
Hello, I just don't get this and would appreciate if someone could write a line or two: I'm trying to build this package and it stops installing after I add the following to the NAMESPACES file: >importFrom(gsubfn,strapply) The error during the package test is: Error in MyPackage::MyFunction : package 'MyPackage' has no name space and is not on the search path Calls:
2008 Aug 12
2
perl expression question
I have a string such as fileName<-"Agg.20.20.20-all-01". All I want to do is pull the "20.20.20" and the "all" as strings. Obviously, they aren't always those values. The "20.20.20" can be "30.30.30" but it's always after the . which is next to the second g in Agg and it's always the same length. The all might not always be