Need help with R studio. Code is to pull data from todays date plus 5 years 4 months from now. I am missing the last 3 months of data. Can you please help? Shawn Miller | Aquatic Biologist II | Assessment Section Environmental Protection | Clean Water Rachel Carson State Office Building 400 Market Street | Harrisburg, PA 17101 Phone: 717.772.2185 | Fax: 717.772.3249 www.depweb.state.pa.us<https://webmail.state.pa.us/OWA/redir.aspx?C=t4jGxr3_mkC5mWY30vM0D8N-9RdJ8s9IgIGFizoEzsd1aNOJaDwGjjpEh4RqLFX24CIJXV9M2ic.&URL=http%3a%2f%2fwww.depweb.state.pa.us%2f> [[alternative HTML version deleted]]
You need Santa Claus not r-help. You haven't given us a fraction of the information we would need to help. You don't show us your code. You don't tell us where the information is coming from except "today's date." You don't tell us what data you want. You don't seem to know the difference between R and R-Studio. ---------------------------------------- David L Carlson Department of Anthropology Texas A&M University College Station, TX 77843-4352 -----Original Message----- From: R-help <r-help-bounces at r-project.org> On Behalf Of Miller, Shawn Sent: Thursday, December 13, 2018 11:29 AM To: R-help at r-project.org Subject: [R] help Need help with R studio. Code is to pull data from todays date plus 5 years 4 months from now. I am missing the last 3 months of data. Can you please help? Shawn Miller | Aquatic Biologist II | Assessment Section Environmental Protection | Clean Water Rachel Carson State Office Building 400 Market Street | Harrisburg, PA 17101 Phone: 717.772.2185 | Fax: 717.772.3249 www.depweb.state.pa.us<https://webmail.state.pa.us/OWA/redir.aspx?C=t4jGxr3_mkC5mWY30vM0D8N-9RdJ8s9IgIGFizoEzsd1aNOJaDwGjjpEh4RqLFX24CIJXV9M2ic.&URL=http%3a%2f%2fwww.depweb.state.pa.us%2f> [[alternative HTML version deleted]] ______________________________________________ R-help at r-project.org mailing list -- To UNSUBSCRIBE and more, see https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide http://www.R-project.org/posting-guide.html and provide commented, minimal, self-contained, reproducible code.
Use Reply-All to keep the thread on the list. If the most recent WQN SIS Data Download table is 14June2018_WQN_Data_Download, does it contain any data for the months July, August, September, October, November, December? After extracting the data you create the file "Anti-deg_requests/test.xlsx". Have you confirmed that all of the dates you need are present in that file? ---------------------------------------- David L Carlson Department of Anthropology Texas A&M University College Station, TX 77843-4352 -----Original Message----- From: Miller, Shawn <shawnmille at pa.gov> Sent: Thursday, December 13, 2018 1:15 PM To: David L Carlson <dcarlson at tamu.edu> Subject: RE: [External] RE: help setwd("c:/users/shawnmille/Desktop/Anti-deg_requests/") library(plyr) library(reshape2) library(RODBC) library(dplyr) ## MAKE SURE YOU ARE USING THE 32-bit VERSION OF R (See Tools>Global Options>General in RStudio to set this) ## READING DATA FROM EXISTING DATABASES (EXAMPLE: ACCESS 2010) chan <- odbcConnectAccess("WQN Datadumps.mdb") # Show list of tables in database sqlTables(chan,tableType='TABLE') # Fetch the most recent WQN SIS Data Download table, this takes a couple minutes allwqn <- sqlFetch(chan,'14June2018_WQN_Data_Download',stringsAsFactors=FALSE) names(allwqn) ## subset to the desired fields and values, choose your WQN by modifying command wqn.specific <- subset(allwqn,MONITORING_POINT_ALIAS_ID =='WQN0735',select=c('DATE_COLLECTED','MONITORING_POINT_ALIAS_ID','QUALITY_ASSURANCE_TYPE_DESC', 'TEST_CODE','TEST_SHORT_DESC','READING_INDICATOR_CODE','FINAL_AMOUNT','ABBREVIATION')) ###VIEW ENTIRE DATASET, this coding can be used to view any created datasets in excel##### library(xlsx) write.xlsx(wqn.specific,"c:/users/shawnmille/Desktop/Anti-deg_requests/test.xlsx") ###VIEW ENTIRE DATASET##### # FOR ACTIVE WQN STATION GET TODAY'S DATE and date of 5 years ago + 4 months or put in last date in command# today <- Sys.Date() TODAY.5<-today-1949 ## Select data where DATA_COLLECTED >= TODAY.5## wqn.specific$DATE_COLLECTED <- as.Date(wqn.specific$DATE_COLLECTED ,"%m/%d/%y") wqn.last5yrs <- subset(wqn.specific,DATE_COLLECTED >= TODAY.5, select=c('DATE_COLLECTED','MONITORING_POINT_ALIAS_ID','QUALITY_ASSURANCE_TYPE_DESC','TEST_CODE','TEST_SHORT_DESC','READING_INDICATOR_CODE','FINAL_AMOUNT','ABBREVIATION')) #OR FOR INACTIVE WQN STATION ENTER THE LAST DATE OF RECORD BELOW AND GET PAST 5 YEARS OF DATA IN SUBSET# lastdate <- as.Date("11/01/2010", format = "%m/%d/%Y") firstdate <- lastdate-1949 wqn.specific$DATE_COLLECTED <- as.Date(wqn.specific$DATE_COLLECTED ,"%m/%d/%y") wqn.last5yrs <- subset(wqn.specific,DATE_COLLECTED >= firstdate & DATE_COLLECTED <= lastdate, select=c('DATE_COLLECTED','MONITORING_POINT_ALIAS_ID','QUALITY_ASSURANCE_TYPE_DESC','TEST_CODE','TEST_SHORT_DESC','READING_INDICATOR_CODE','FINAL_AMOUNT','ABBREVIATION')) #### Delete Blanks and QA column ##### wqn.removeblanks <-subset(wqn.last5yrs,QUALITY_ASSURANCE_TYPE_DESC=="Duplicate"|is.na(QUALITY_ASSURANCE_TYPE_DESC),select=c('DATE_COLLECTED','MONITORING_POINT_ALIAS_ID', 'TEST_CODE','TEST_SHORT_DESC','READING_INDICATOR_CODE','FINAL_AMOUNT','ABBREVIATION')) #### Delete Null Results#### wqn.removenull <- subset(wqn.removeblanks,FINAL_AMOUNT!="NA",select=c('DATE_COLLECTED','MONITORING_POINT_ALIAS_ID','TEST_CODE','TEST_SHORT_DESC','READING_INDICATOR_CODE','FINAL_AMOUNT','ABBREVIATION')) ### Remove non-relevant parameters #### wqn.removeparm <- subset(wqn.removenull,TEST_CODE!= "571"&TEST_CODE!="572"& TEST_CODE!="573"& TEST_CODE!="543"& TEST_CODE!="561"& TEST_CODE!="546" & TEST_CODE!="595"& TEST_CODE!="596"& TEST_CODE!="597"& TEST_CODE!="598" & TEST_CODE!="599"& TEST_CODE!="600"& TEST_CODE!="71946"& TEST_CODE!="71940" & TEST_CODE!="587"& TEST_CODE!="593"& TEST_CODE!="71939"& TEST_CODE!="71937" & TEST_CODE!="574"& TEST_CODE!="549"& TEST_CODE!="99014"& TEST_CODE!="547" & TEST_CODE!="298"& TEST_CODE!="551"& TEST_CODE!="71930"& TEST_CODE!="70508" & TEST_CODE!="564"& TEST_CODE!="709"& TEST_CODE!="111"& TEST_CODE!="592" & TEST_CODE!="MMTECMF"& TEST_CODE!="588"& TEST_CODE!="589"& TEST_CODE!="590" & TEST_CODE!="594"& TEST_CODE!="71936"& TEST_CODE!="71945"& TEST_CODE!="F00061" & TEST_CODE!="71947"& TEST_CODE!="555", select=c('DATE_COLLECTED','MONITORING_POINT_ALIAS_ID','TEST_CODE','TEST_SHORT_DESC','READING_INDICATOR_CODE','FINAL_AMOUNT','ABBREVIATION')) ### Combine test codes that are measuring the same analyte, format testcode to include first 5 digits### wqn.removeparm$TEST_CODE<-substr(wqn.removeparm$TEST_CODE,1,5) ######Creates a table (wqn.det) with list of max detection limits for each parameter,FOR EACH TEST_CODE BRING ALL < TO THAT VALUE############ n.det <- subset(wqn.removeparm,READING_INDICATOR_CODE=='<') wqn.det<- aggregate(n.det$FINAL_AMOUNT , by=list(n.det$MONITORING_POINT_ALIAS_ID,n.det$TEST_CODE),FUN=max) colnames(wqn.det) <- c("MONITORING_POINT_ALIAS_ID","TEST_CODE","DET_LIMIT") wqn.mer<-merge(wqn.removeparm,wqn.det, by=c("MONITORING_POINT_ALIAS_ID","TEST_CODE"),all=TRUE) wqn.mer$Amount<- ifelse(is.na(wqn.mer$'READING_INDICATOR_CODE'),wqn.mer$FINAL_AMOUNT*1,wqn.mer$DET_LIMIT*1) ######DIVIDE '<' BY HALF ##### wqn.mer$Result <- ifelse(is.na(wqn.mer$'READING_INDICATOR_CODE'),wqn.mer$Amount*1,wqn.mer$Amount/2) ###### AVERAGE AND GROUP BY TO REMOVE DUPLICATES ######## wqn.agg<-aggregate(wqn.mer$Result, by=list(MONITORING_POINT_ALIAS_ID=wqn.mer$'MONITORING_POINT_ALIAS_ID',DATE_COLLECTED=wqn.mer$'DATE_COLLECTED',TEST_CODE=wqn.mer$'TEST_CODE'),data=wqn.mer, FUN="mean",na.rm=TRUE) colnames(wqn.agg) <- c("MONITORING_POINT_ALIAS_ID","DATE_COLLECTED","TEST_CODE","Result") ###### Calculate Median and 95% CI, alpha/2 for tailed2 (pH)..similar to ci.median function in asbio package, using alpha for funtion "tailed1"(other parameters)### tailed2<-function(x){ n <- nrow(as.matrix(x)) L <- qbinom(.025,n,0.5) U <- n-L+1 order.x <-sort(x) lower <- (order.x[L]) upper <-order.x[n - L + 1] median <- median(x) coverage <- 1-(2*pbinom(qbinom(.025,n,.5)-1,n,0.5)) y<-list(c(lower, median, upper,n,coverage)) } tailed1<-function(x){ n <- nrow(as.matrix(x)) L <- qbinom(.05,n,0.5) U <- n-L+1 order.x <-sort(x) lower <- (order.x[L]) upper <-order.x[n - L+1] median <- median(x) coverage <- 1-(2*pbinom(qbinom(.05,n,.5)-1,n,0.5)) y<-list(c(lower, median, upper,n,coverage)) } ###Create two data sets, one with all the parameters except pH and one with just pH##### wqn.ph <- subset(wqn.agg,TEST_CODE=='00403'|TEST_CODE=="F0040",select=c("MONITORING_POINT_ALIAS_ID","DATE_COLLECTED","TEST_CODE","Result")) wqn.noph <-subset(wqn.agg,TEST_CODE!='00403'& TEST_CODE!="F0040",select=c("MONITORING_POINT_ALIAS_ID","DATE_COLLECTED","TEST_CODE","Result")) ###Run tailed1 function on all data except for pH### stats<-aggregate(wqn.noph$Result, by=list(wqn.noph$MONITORING_POINT_ALIAS_ID,wqn.noph$TEST_CODE), FUN=tailed1) colnames(stats) <- c("MONITORING_POINT_ALIAS_ID","TEST_CODE","x") stats$lower <-unlist(lapply(stats$x, '[[', 1)) stats$median<-unlist(lapply(stats$x, '[[', 2)) stats$upper <-unlist(lapply(stats$x, '[[', 3)) stats$n <-unlist(lapply(stats$x, '[[', 4)) stats1 <- stats[order(stats$"MONITORING_POINT_ALIAS_ID",stats$"TEST_CODE"),] ###Run tailed2 function on pH (lab and field) data#### ph.stats<-aggregate(wqn.ph$Result, by=list(wqn.ph$MONITORING_POINT_ALIAS_ID,wqn.ph$TEST_CODE), FUN=tailed2) colnames(ph.stats) <- c("MONITORING_POINT_ALIAS_ID","TEST_CODE","x") ph.stats$lower <-unlist(lapply(ph.stats$x, '[[', 1)) ph.stats$median<-unlist(lapply(ph.stats$x, '[[', 2)) ph.stats$upper <-unlist(lapply(ph.stats$x, '[[', 3)) ph.stats$n <-unlist(lapply(ph.stats$x, '[[', 4)) ph.stats1 <- ph.stats[order(ph.stats$"MONITORING_POINT_ALIAS_ID",ph.stats$"TEST_CODE"),] ###Combine datasets vertically### stats2 <- rbind(stats1,ph.stats1) ####PeriodofRecordTable#### library(dplyr) my.dt<-(format(as.Date(wqn.removeparm$"DATE_COLLECTED"),"%m/%d/%Y")) my.dt1<-aggregate(my.dt, by=list(wqn.removeparm$MONITORING_POINT_ALIAS_ID,wqn.removeparm$TEST_CODE), FUN=first) colnames(my.dt1) <- c("MONITORING_POINT_ALIAS_ID","TEST_CODE","FirstDate") my.dtlast<-aggregate(my.dt, by=list(wqn.removeparm$MONITORING_POINT_ALIAS_ID,wqn.removeparm$TEST_CODE), FUN=last) colnames(my.dtlast) <- c("MONITORING_POINT_ALIAS_ID","TEST_CODE","LastDate") ####Merging Tables Dates and Stats#################### all.dt <- merge(my.dt1,my.dtlast,by=c("MONITORING_POINT_ALIAS_ID","TEST_CODE")) dt.stats<-merge(all.dt,stats2,by=c("MONITORING_POINT_ALIAS_ID","TEST_CODE")) ####Calculate Period Of record in Table, Old coding 2 lines: dt.stats$pdofrecord<-(mdy(dt.stats$"LastDate"))-(mdy(dt.stats$"FirstDate")) ###dt.stats$pd<-round(dt.stats$pdofrecord/31563000, digits=1), lubridate is glitchy test new code on this step more###### dt.stats$pd <- round((as.Date(dt.stats$LastDate,format = "%m/%d/%Y")-as.Date(dt.stats$FirstDate,format = "%m/%d/%Y"))/365,digits=1) ####Retrieve Test Descriptions and Merge###### t1 <- subset(dt.stats,select=c('MONITORING_POINT_ALIAS_ID','TEST_CODE','FirstDate','LastDate','pd','lower','median','upper','n')) t2<-subset(wqn.removeparm,select=c('TEST_CODE','TEST_SHORT_DESC','ABBREVIATION')) t3 <- unique( t2 ) t4<-merge(t1,t3, by='TEST_CODE',all.t3=TRUE) ######ReorderColumns,sort, rename columns####### t5 <- t4[ ,c(2,1,10,3,4,5,9,6,7,8,11)] t6 <- t5[order(t5$"MONITORING_POINT_ALIAS_ID",t5$"TEST_SHORT_DESC"),] names(t6) ####This does not need to be run. It produces an error. You still get the correct results. library(plyr) library(dplyr) library(reshape2) final.tab<-rename(t6, c("FirstDate"="FIRST_DATE", "LastDate"="LAST_DATE","pd"="PERIOD_OF_RECORD(yrs)","n"="SAMPLE_SIZE","lower" ="L_95_CI","median"="MEDIAN_","upper"="U_95_CI","ABBREVIATION"="UNITS")) ####BRING VALUES UP TO DETECTION LIMITS merge with wqn.det###### less<-function(x){ sprintf("< %3.2f", x) } wqn.final <-merge(t6,wqn.det,by=c("MONITORING_POINT_ALIAS_ID","TEST_CODE"),all=TRUE) wqn.final$DET_LIMIT[is.na(wqn.final$DET_LIMIT)] <- 0 wqn.final$LOW_95_CL <- ifelse(wqn.final$lower < wqn.final$DET_LIM,less(wqn.final$DET_LIM*1),wqn.final$lower*1) wqn.final$MEDIAN <- ifelse(wqn.final$median < wqn.final$DET_LIM,less(wqn.final$DET_LIM*1),wqn.final$median*1) wqn.final$UPP_95_CL <- ifelse(wqn.final$upper < wqn.final$DET_LIM,less(wqn.final$DET_LIM*1),wqn.final$upper*1) ####ONLY REPORT MINIMUM FOR ALKALINITY, PH, DO, TEMP, SPECIFIC cONDUCTANCE###### wqn.final$LOWER_95_CL <- ifelse(wqn.final$TEST_CODE=="00403"|wqn.final$TEST_CODE=="00410"|wqn.final$TEST_CODE=="F0040"| wqn.final$TEST_CODE=="F0030"|wqn.final$TEST_CODE=="F0043",wqn.final$LOW_95_CL,"NA") wqn.final$UPPER_95_CL <- ifelse(wqn.final$TEST_CODE=="00410"|wqn.final$TEST_CODE=="F0043"| wqn.final$TEST_CODE=="F0030","NA",wqn.final$UPP_95_CL) names(wqn.final) final.1<-subset(wqn.final, select=c('MONITORING_POINT_ALIAS_ID','TEST_CODE','TEST_SHORT_DESC','FirstDate', 'LastDate', 'pd',"n","LOWER_95_CL",'MEDIAN', 'UPPER_95_CL','ABBREVIATION')) final.2 <- final.1[order(final.1$"MONITORING_POINT_ALIAS_ID",final.1$"TEST_SHORT_DESC"),] ###VIEW IN EXCEL## library(xlsx) write.xlsx(final.2,"c:/users/shawnmille/Desktop/Anti-deg_requests/final.2.xlsx") -----Original Message----- From: David L Carlson [mailto:dcarlson at tamu.edu] Sent: Thursday, December 13, 2018 2:12 PM To: Miller, Shawn <shawnmille at pa.gov>; R-help at r-project.org Subject: [External] RE: help ATTENTION: This email message is from an external sender. Do not open links or attachments from unknown sources. To report suspicious email, forward the message as an attachment to CWOPA_SPAM at pa.gov. You need Santa Claus not r-help. You haven't given us a fraction of the information we would need to help. You don't show us your code. You don't tell us where the information is coming from except "today's date." You don't tell us what data you want. You don't seem to know the difference between R and R-Studio. ---------------------------------------- David L Carlson Department of Anthropology Texas A&M University College Station, TX 77843-4352 -----Original Message----- From: R-help <r-help-bounces at r-project.org> On Behalf Of Miller, Shawn Sent: Thursday, December 13, 2018 11:29 AM To: R-help at r-project.org Subject: [R] help Need help with R studio. Code is to pull data from todays date plus 5 years 4 months from now. I am missing the last 3 months of data. Can you please help? Shawn Miller | Aquatic Biologist II | Assessment Section Environmental Protection | Clean Water Rachel Carson State Office Building 400 Market Street | Harrisburg, PA 17101 Phone: 717.772.2185 | Fax: 717.772.3249 https://na01.safelinks.protection.outlook.com/?url=www.depweb.state.pa.us&data=02%7C01%7Cshawnmille%40pa.gov%7Cdfe052e5375647a1c32f08d6612eec13%7C418e284101284dd59b6c47fc5a9a1bde%7C1%7C1%7C636803251493279248&sdata=ScL5NIrFc1A4g9wKVAYxhZ6Hrwnb7qvgjijh2WHEENY%3D&reserved=0<https://na01.safelinks.protection.outlook.com/?url=https%3A%2F%2Fwebmail.state.pa.us%2FOWA%2Fredir.aspx%3FC%3Dt4jGxr3_mkC5mWY30vM0D8N-9RdJ8s9IgIGFizoEzsd1aNOJaDwGjjpEh4RqLFX24CIJXV9M2ic.%26URL%3Dhttp%253a%252f%252fwww.depweb.state.pa.us%252f&data=02%7C01%7Cshawnmille%40pa.gov%7Cdfe052e5375647a1c32f08d6612eec13%7C418e284101284dd59b6c47fc5a9a1bde%7C1%7C1%7C636803251493279248&sdata=H6VzxlE9wQCOszIoDOZXOoI5%2BZmsu2qVfP1Y%2BEU50SU%3D&reserved=0> [[alternative HTML version deleted]] ______________________________________________ R-help at r-project.org mailing list -- To UNSUBSCRIBE and more, see https://na01.safelinks.protection.outlook.com/?url=https%3A%2F%2Fstat.ethz.ch%2Fmailman%2Flistinfo%2Fr-help&data=02%7C01%7Cshawnmille%40pa.gov%7Cdfe052e5375647a1c32f08d6612eec13%7C418e284101284dd59b6c47fc5a9a1bde%7C1%7C1%7C636803251493279248&sdata=WM4AQs45%2FgKgmoxl4bS%2B5sTSaWYhCt0YhmkVSlwqgeY%3D&reserved=0 PLEASE do read the posting guide https://na01.safelinks.protection.outlook.com/?url=http%3A%2F%2Fwww.R-project.org%2Fposting-guide.html&data=02%7C01%7Cshawnmille%40pa.gov%7Cdfe052e5375647a1c32f08d6612eec13%7C418e284101284dd59b6c47fc5a9a1bde%7C1%7C1%7C636803251493279248&sdata=BSVgu1ome5ob4rGe%2Bck3hcInxJn6tF0N%2F1CBJezIKG4%3D&reserved=0 and provide commented, minimal, self-contained, reproducible code.
Yes I confirmed all the dates I need are in the test.xlsx table. The final.2.xlsx table does not have the dates for July, August, September. -----Original Message----- From: David L Carlson [mailto:dcarlson at tamu.edu] Sent: Thursday, December 13, 2018 2:33 PM To: Miller, Shawn <shawnmille at pa.gov> Cc: R-help at r-project.org Subject: RE: [External] RE: help Use Reply-All to keep the thread on the list. If the most recent WQN SIS Data Download table is 14June2018_WQN_Data_Download, does it contain any data for the months July, August, September, October, November, December? After extracting the data you create the file "Anti-deg_requests/test.xlsx". Have you confirmed that all of the dates you need are present in that file? ---------------------------------------- David L Carlson Department of Anthropology Texas A&M University College Station, TX 77843-4352 -----Original Message----- From: Miller, Shawn <shawnmille at pa.gov> Sent: Thursday, December 13, 2018 1:15 PM To: David L Carlson <dcarlson at tamu.edu> Subject: RE: [External] RE: help setwd("c:/users/shawnmille/Desktop/Anti-deg_requests/") library(plyr) library(reshape2) library(RODBC) library(dplyr) ## MAKE SURE YOU ARE USING THE 32-bit VERSION OF R (See Tools>Global Options>General in RStudio to set this) ## READING DATA FROM EXISTING DATABASES (EXAMPLE: ACCESS 2010) chan <- odbcConnectAccess("WQN Datadumps.mdb") # Show list of tables in database sqlTables(chan,tableType='TABLE') # Fetch the most recent WQN SIS Data Download table, this takes a couple minutes allwqn <- sqlFetch(chan,'14June2018_WQN_Data_Download',stringsAsFactors=FALSE) names(allwqn) ## subset to the desired fields and values, choose your WQN by modifying command wqn.specific <- subset(allwqn,MONITORING_POINT_ALIAS_ID =='WQN0735',select=c('DATE_COLLECTED','MONITORING_POINT_ALIAS_ID','QUALITY_ASSURANCE_TYPE_DESC', 'TEST_CODE','TEST_SHORT_DESC','READING_INDICATOR_CODE','FINAL_AMOUNT','ABBREVIATION')) ###VIEW ENTIRE DATASET, this coding can be used to view any created datasets in excel##### library(xlsx) write.xlsx(wqn.specific,"c:/users/shawnmille/Desktop/Anti-deg_requests/test.xlsx") ###VIEW ENTIRE DATASET##### # FOR ACTIVE WQN STATION GET TODAY'S DATE and date of 5 years ago + 4 months or put in last date in command# today <- Sys.Date() TODAY.5<-today-1949 ## Select data where DATA_COLLECTED >= TODAY.5## wqn.specific$DATE_COLLECTED <- as.Date(wqn.specific$DATE_COLLECTED ,"%m/%d/%y") wqn.last5yrs <- subset(wqn.specific,DATE_COLLECTED >= TODAY.5, select=c('DATE_COLLECTED','MONITORING_POINT_ALIAS_ID','QUALITY_ASSURANCE_TYPE_DESC','TEST_CODE','TEST_SHORT_DESC','READING_INDICATOR_CODE','FINAL_AMOUNT','ABBREVIATION')) #OR FOR INACTIVE WQN STATION ENTER THE LAST DATE OF RECORD BELOW AND GET PAST 5 YEARS OF DATA IN SUBSET# lastdate <- as.Date("11/01/2010", format = "%m/%d/%Y") firstdate <- lastdate-1949 wqn.specific$DATE_COLLECTED <- as.Date(wqn.specific$DATE_COLLECTED ,"%m/%d/%y") wqn.last5yrs <- subset(wqn.specific,DATE_COLLECTED >= firstdate & DATE_COLLECTED <= lastdate, select=c('DATE_COLLECTED','MONITORING_POINT_ALIAS_ID','QUALITY_ASSURANCE_TYPE_DESC','TEST_CODE','TEST_SHORT_DESC','READING_INDICATOR_CODE','FINAL_AMOUNT','ABBREVIATION')) #### Delete Blanks and QA column ##### wqn.removeblanks <-subset(wqn.last5yrs,QUALITY_ASSURANCE_TYPE_DESC=="Duplicate"|is.na(QUALITY_ASSURANCE_TYPE_DESC),select=c('DATE_COLLECTED','MONITORING_POINT_ALIAS_ID', 'TEST_CODE','TEST_SHORT_DESC','READING_INDICATOR_CODE','FINAL_AMOUNT','ABBREVIATION')) #### Delete Null Results#### wqn.removenull <- subset(wqn.removeblanks,FINAL_AMOUNT!="NA",select=c('DATE_COLLECTED','MONITORING_POINT_ALIAS_ID','TEST_CODE','TEST_SHORT_DESC','READING_INDICATOR_CODE','FINAL_AMOUNT','ABBREVIATION')) ### Remove non-relevant parameters #### wqn.removeparm <- subset(wqn.removenull,TEST_CODE!= "571"&TEST_CODE!="572"& TEST_CODE!="573"& TEST_CODE!="543"& TEST_CODE!="561"& TEST_CODE!="546" & TEST_CODE!="595"& TEST_CODE!="596"& TEST_CODE!="597"& TEST_CODE!="598" & TEST_CODE!="599"& TEST_CODE!="600"& TEST_CODE!="71946"& TEST_CODE!="71940" & TEST_CODE!="587"& TEST_CODE!="593"& TEST_CODE!="71939"& TEST_CODE!="71937" & TEST_CODE!="574"& TEST_CODE!="549"& TEST_CODE!="99014"& TEST_CODE!="547" & TEST_CODE!="298"& TEST_CODE!="551"& TEST_CODE!="71930"& TEST_CODE!="70508" & TEST_CODE!="564"& TEST_CODE!="709"& TEST_CODE!="111"& TEST_CODE!="592" & TEST_CODE!="MMTECMF"& TEST_CODE!="588"& TEST_CODE!="589"& TEST_CODE!="590" & TEST_CODE!="594"& TEST_CODE!="71936"& TEST_CODE!="71945"& TEST_CODE!="F00061" & TEST_CODE!="71947"& TEST_CODE!="555", select=c('DATE_COLLECTED','MONITORING_POINT_ALIAS_ID','TEST_CODE','TEST_SHORT_DESC','READING_INDICATOR_CODE','FINAL_AMOUNT','ABBREVIATION')) ### Combine test codes that are measuring the same analyte, format testcode to include first 5 digits### wqn.removeparm$TEST_CODE<-substr(wqn.removeparm$TEST_CODE,1,5) ######Creates a table (wqn.det) with list of max detection limits for each parameter,FOR EACH TEST_CODE BRING ALL < TO THAT VALUE############ n.det <- subset(wqn.removeparm,READING_INDICATOR_CODE=='<') wqn.det<- aggregate(n.det$FINAL_AMOUNT , by=list(n.det$MONITORING_POINT_ALIAS_ID,n.det$TEST_CODE),FUN=max) colnames(wqn.det) <- c("MONITORING_POINT_ALIAS_ID","TEST_CODE","DET_LIMIT") wqn.mer<-merge(wqn.removeparm,wqn.det, by=c("MONITORING_POINT_ALIAS_ID","TEST_CODE"),all=TRUE) wqn.mer$Amount<- ifelse(is.na(wqn.mer$'READING_INDICATOR_CODE'),wqn.mer$FINAL_AMOUNT*1,wqn.mer$DET_LIMIT*1) ######DIVIDE '<' BY HALF ##### wqn.mer$Result <- ifelse(is.na(wqn.mer$'READING_INDICATOR_CODE'),wqn.mer$Amount*1,wqn.mer$Amount/2) ###### AVERAGE AND GROUP BY TO REMOVE DUPLICATES ######## wqn.agg<-aggregate(wqn.mer$Result, by=list(MONITORING_POINT_ALIAS_ID=wqn.mer$'MONITORING_POINT_ALIAS_ID',DATE_COLLECTED=wqn.mer$'DATE_COLLECTED',TEST_CODE=wqn.mer$'TEST_CODE'),data=wqn.mer, FUN="mean",na.rm=TRUE) colnames(wqn.agg) <- c("MONITORING_POINT_ALIAS_ID","DATE_COLLECTED","TEST_CODE","Result") ###### Calculate Median and 95% CI, alpha/2 for tailed2 (pH)..similar to ci.median function in asbio package, using alpha for funtion "tailed1"(other parameters)### tailed2<-function(x){ n <- nrow(as.matrix(x)) L <- qbinom(.025,n,0.5) U <- n-L+1 order.x <-sort(x) lower <- (order.x[L]) upper <-order.x[n - L + 1] median <- median(x) coverage <- 1-(2*pbinom(qbinom(.025,n,.5)-1,n,0.5)) y<-list(c(lower, median, upper,n,coverage)) } tailed1<-function(x){ n <- nrow(as.matrix(x)) L <- qbinom(.05,n,0.5) U <- n-L+1 order.x <-sort(x) lower <- (order.x[L]) upper <-order.x[n - L+1] median <- median(x) coverage <- 1-(2*pbinom(qbinom(.05,n,.5)-1,n,0.5)) y<-list(c(lower, median, upper,n,coverage)) } ###Create two data sets, one with all the parameters except pH and one with just pH##### wqn.ph <- subset(wqn.agg,TEST_CODE=='00403'|TEST_CODE=="F0040",select=c("MONITORING_POINT_ALIAS_ID","DATE_COLLECTED","TEST_CODE","Result")) wqn.noph <-subset(wqn.agg,TEST_CODE!='00403'& TEST_CODE!="F0040",select=c("MONITORING_POINT_ALIAS_ID","DATE_COLLECTED","TEST_CODE","Result")) ###Run tailed1 function on all data except for pH### stats<-aggregate(wqn.noph$Result, by=list(wqn.noph$MONITORING_POINT_ALIAS_ID,wqn.noph$TEST_CODE), FUN=tailed1) colnames(stats) <- c("MONITORING_POINT_ALIAS_ID","TEST_CODE","x") stats$lower <-unlist(lapply(stats$x, '[[', 1)) stats$median<-unlist(lapply(stats$x, '[[', 2)) stats$upper <-unlist(lapply(stats$x, '[[', 3)) stats$n <-unlist(lapply(stats$x, '[[', 4)) stats1 <- stats[order(stats$"MONITORING_POINT_ALIAS_ID",stats$"TEST_CODE"),] ###Run tailed2 function on pH (lab and field) data#### ph.stats<-aggregate(wqn.ph$Result, by=list(wqn.ph$MONITORING_POINT_ALIAS_ID,wqn.ph$TEST_CODE), FUN=tailed2) colnames(ph.stats) <- c("MONITORING_POINT_ALIAS_ID","TEST_CODE","x") ph.stats$lower <-unlist(lapply(ph.stats$x, '[[', 1)) ph.stats$median<-unlist(lapply(ph.stats$x, '[[', 2)) ph.stats$upper <-unlist(lapply(ph.stats$x, '[[', 3)) ph.stats$n <-unlist(lapply(ph.stats$x, '[[', 4)) ph.stats1 <- ph.stats[order(ph.stats$"MONITORING_POINT_ALIAS_ID",ph.stats$"TEST_CODE"),] ###Combine datasets vertically### stats2 <- rbind(stats1,ph.stats1) ####PeriodofRecordTable#### library(dplyr) my.dt<-(format(as.Date(wqn.removeparm$"DATE_COLLECTED"),"%m/%d/%Y")) my.dt1<-aggregate(my.dt, by=list(wqn.removeparm$MONITORING_POINT_ALIAS_ID,wqn.removeparm$TEST_CODE), FUN=first) colnames(my.dt1) <- c("MONITORING_POINT_ALIAS_ID","TEST_CODE","FirstDate") my.dtlast<-aggregate(my.dt, by=list(wqn.removeparm$MONITORING_POINT_ALIAS_ID,wqn.removeparm$TEST_CODE), FUN=last) colnames(my.dtlast) <- c("MONITORING_POINT_ALIAS_ID","TEST_CODE","LastDate") ####Merging Tables Dates and Stats#################### all.dt <- merge(my.dt1,my.dtlast,by=c("MONITORING_POINT_ALIAS_ID","TEST_CODE")) dt.stats<-merge(all.dt,stats2,by=c("MONITORING_POINT_ALIAS_ID","TEST_CODE")) ####Calculate Period Of record in Table, Old coding 2 lines: dt.stats$pdofrecord<-(mdy(dt.stats$"LastDate"))-(mdy(dt.stats$"FirstDate")) ###dt.stats$pd<-round(dt.stats$pdofrecord/31563000, digits=1), lubridate is glitchy test new code on this step more###### dt.stats$pd <- round((as.Date(dt.stats$LastDate,format = "%m/%d/%Y")-as.Date(dt.stats$FirstDate,format = "%m/%d/%Y"))/365,digits=1) ####Retrieve Test Descriptions and Merge###### t1 <- subset(dt.stats,select=c('MONITORING_POINT_ALIAS_ID','TEST_CODE','FirstDate','LastDate','pd','lower','median','upper','n')) t2<-subset(wqn.removeparm,select=c('TEST_CODE','TEST_SHORT_DESC','ABBREVIATION')) t3 <- unique( t2 ) t4<-merge(t1,t3, by='TEST_CODE',all.t3=TRUE) ######ReorderColumns,sort, rename columns####### t5 <- t4[ ,c(2,1,10,3,4,5,9,6,7,8,11)] t6 <- t5[order(t5$"MONITORING_POINT_ALIAS_ID",t5$"TEST_SHORT_DESC"),] names(t6) ####This does not need to be run. It produces an error. You still get the correct results. library(plyr) library(dplyr) library(reshape2) final.tab<-rename(t6, c("FirstDate"="FIRST_DATE", "LastDate"="LAST_DATE","pd"="PERIOD_OF_RECORD(yrs)","n"="SAMPLE_SIZE","lower" ="L_95_CI","median"="MEDIAN_","upper"="U_95_CI","ABBREVIATION"="UNITS")) ####BRING VALUES UP TO DETECTION LIMITS merge with wqn.det###### less<-function(x){ sprintf("< %3.2f", x) } wqn.final <-merge(t6,wqn.det,by=c("MONITORING_POINT_ALIAS_ID","TEST_CODE"),all=TRUE) wqn.final$DET_LIMIT[is.na(wqn.final$DET_LIMIT)] <- 0 wqn.final$LOW_95_CL <- ifelse(wqn.final$lower < wqn.final$DET_LIM,less(wqn.final$DET_LIM*1),wqn.final$lower*1) wqn.final$MEDIAN <- ifelse(wqn.final$median < wqn.final$DET_LIM,less(wqn.final$DET_LIM*1),wqn.final$median*1) wqn.final$UPP_95_CL <- ifelse(wqn.final$upper < wqn.final$DET_LIM,less(wqn.final$DET_LIM*1),wqn.final$upper*1) ####ONLY REPORT MINIMUM FOR ALKALINITY, PH, DO, TEMP, SPECIFIC cONDUCTANCE###### wqn.final$LOWER_95_CL <- ifelse(wqn.final$TEST_CODE=="00403"|wqn.final$TEST_CODE=="00410"|wqn.final$TEST_CODE=="F0040"| wqn.final$TEST_CODE=="F0030"|wqn.final$TEST_CODE=="F0043",wqn.final$LOW_95_CL,"NA") wqn.final$UPPER_95_CL <- ifelse(wqn.final$TEST_CODE=="00410"|wqn.final$TEST_CODE=="F0043"| wqn.final$TEST_CODE=="F0030","NA",wqn.final$UPP_95_CL) names(wqn.final) final.1<-subset(wqn.final, select=c('MONITORING_POINT_ALIAS_ID','TEST_CODE','TEST_SHORT_DESC','FirstDate', 'LastDate', 'pd',"n","LOWER_95_CL",'MEDIAN', 'UPPER_95_CL','ABBREVIATION')) final.2 <- final.1[order(final.1$"MONITORING_POINT_ALIAS_ID",final.1$"TEST_SHORT_DESC"),] ###VIEW IN EXCEL## library(xlsx) write.xlsx(final.2,"c:/users/shawnmille/Desktop/Anti-deg_requests/final.2.xlsx") -----Original Message----- From: David L Carlson [mailto:dcarlson at tamu.edu] Sent: Thursday, December 13, 2018 2:12 PM To: Miller, Shawn <shawnmille at pa.gov>; R-help at r-project.org Subject: [External] RE: help ATTENTION: This email message is from an external sender. Do not open links or attachments from unknown sources. To report suspicious email, forward the message as an attachment to CWOPA_SPAM at pa.gov. You need Santa Claus not r-help. You haven't given us a fraction of the information we would need to help. You don't show us your code. You don't tell us where the information is coming from except "today's date." You don't tell us what data you want. You don't seem to know the difference between R and R-Studio. ---------------------------------------- David L Carlson Department of Anthropology Texas A&M University College Station, TX 77843-4352 -----Original Message----- From: R-help <r-help-bounces at r-project.org> On Behalf Of Miller, Shawn Sent: Thursday, December 13, 2018 11:29 AM To: R-help at r-project.org Subject: [R] help Need help with R studio. Code is to pull data from todays date plus 5 years 4 months from now. I am missing the last 3 months of data. Can you please help? Shawn Miller | Aquatic Biologist II | Assessment Section Environmental Protection | Clean Water Rachel Carson State Office Building 400 Market Street | Harrisburg, PA 17101 Phone: 717.772.2185 | Fax: 717.772.3249 https://na01.safelinks.protection.outlook.com/?url=www.depweb.state.pa.us&data=02%7C01%7Cshawnmille%40pa.gov%7Ce275c28afad54d99b5d608d66131bec5%7C418e284101284dd59b6c47fc5a9a1bde%7C1%7C0%7C636803263627811233&sdata=SetioLDNtYwPWgXTfgp4XNoH2ptEpbMkVCkYBcBFJAE%3D&reserved=0<https://na01.safelinks.protection.outlook.com/?url=https%3A%2F%2Fwebmail.state.pa.us%2FOWA%2Fredir.aspx%3FC%3Dt4jGxr3_mkC5mWY30vM0D8N-9RdJ8s9IgIGFizoEzsd1aNOJaDwGjjpEh4RqLFX24CIJXV9M2ic.%26URL%3Dhttp%253a%252f%252fwww.depweb.state.pa.us%252f&data=02%7C01%7Cshawnmille%40pa.gov%7Ce275c28afad54d99b5d608d66131bec5%7C418e284101284dd59b6c47fc5a9a1bde%7C1%7C0%7C636803263627811233&sdata=8mPY9gVwTHUVb6EhpRueVUcHQIyTyd7QYYctDdbwBsM%3D&reserved=0> [[alternative HTML version deleted]] ______________________________________________ R-help at r-project.org mailing list -- To UNSUBSCRIBE and more, see https://na01.safelinks.protection.outlook.com/?url=https%3A%2F%2Fstat.ethz.ch%2Fmailman%2Flistinfo%2Fr-help&data=02%7C01%7Cshawnmille%40pa.gov%7Ce275c28afad54d99b5d608d66131bec5%7C418e284101284dd59b6c47fc5a9a1bde%7C1%7C0%7C636803263627811233&sdata=FEb6F5f1w9Y7%2BPPPfHvCkQNVZjNJJjHod%2F0%2BX4IifUw%3D&reserved=0 PLEASE do read the posting guide https://na01.safelinks.protection.outlook.com/?url=http%3A%2F%2Fwww.R-project.org%2Fposting-guide.html&data=02%7C01%7Cshawnmille%40pa.gov%7Ce275c28afad54d99b5d608d66131bec5%7C418e284101284dd59b6c47fc5a9a1bde%7C1%7C0%7C636803263627811233&sdata=ZemT7Y8LV%2FFIi5J46qF1rN48UlM1OxelnJYFhcmvLRc%3D&reserved=0 and provide commented, minimal, self-contained, reproducible code.
"You need Santa Claus not r-help." in response to an unrealistic and poorly posed request for help. Best, Chuck> On Dec 13, 2018, at 11:12 AM, David L Carlson <dcarlson at tamu.edu> wrote: > > You need Santa Claus not r-help. You haven't given us a fraction of the information we would need to help. You don't show us your code. You don't tell us where the information is coming from except "today's date." You don't tell us what data you want. You don't seem to know the difference between R and R-Studio. > > ---------------------------------------- > David L Carlson > Department of Anthropology > Texas A&M University > College Station, TX 77843-4352 > > > -----Original Message----- > From: R-help <r-help-bounces at r-project.org> On Behalf Of Miller, Shawn > Sent: Thursday, December 13, 2018 11:29 AM > To: R-help at r-project.org > Subject: [R] help > > Need help with R studio. Code is to pull data from todays date plus 5 years 4 months from now. I am missing the last 3 months of data. Can you please help? > > Shawn Miller | Aquatic Biologist II | Assessment Section > Environmental Protection | Clean Water > Rachel Carson State Office Building > 400 Market Street | Harrisburg, PA 17101 > Phone: 717.772.2185 | Fax: 717.772.3249 > www.depweb.state.pa.us<https://webmail.state.pa.us/OWA/redir.aspx?C=t4jGxr3_mkC5mWY30vM0D8N-9RdJ8s9IgIGFizoEzsd1aNOJaDwGjjpEh4RqLFX24CIJXV9M2ic.&URL=http%3a%2f%2fwww.depweb.state.pa.us%2f> > > > [[alternative HTML version deleted]] > > ______________________________________________ > R-help at r-project.org mailing list -- To UNSUBSCRIBE and more, see > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide http://www.R-project.org/posting-guide.html > and provide commented, minimal, self-contained, reproducible code. >