similar to: get only little part of html with htmlParse

Displaying 20 results from an estimated 300 matches similar to: "get only little part of html with htmlParse"

2013 Mar 20
1
htmlParse (from XML library) working sporadically in the same code
I am using htmlParse from XML library on a paricular website. Sometimes code fails, sometimes it works, most of the time id doesn't and i cannot see why. The file i am trying to parse is  http://www.londonstockexchange.com/exchange/prices-and-markets/international-markets/indices/home/sp-500.html?page=0 Sometimes the following code works n<-readHTMLTable(htmlParse(url)) But most of the
2012 May 21
1
htmlParse Error
I am trying to parse a webpage using the htmlParse command in XML package as follows: library(XML) u = "http://en.wikipedia.org/wiki/World_population" doc = htmlParse(u) I get the following error: Error in htmlParse(u) : error in creating parser for http://en.wikipedia.org/wiki/World_population I am using a R 2.13.1 (32 bit version) on a 64 bit Windows. (I tried installing it in
2012 Jan 30
1
Getting htmlParse to work with Hebrew? (on windows)
Hello dear R-help mailing list. I wish to be able to have htmlParse work well with Hebrew, but it keeps to scramble the Hebrew text in pages I feed into it. For example: # why can't I parse the Hebrew correctly? library(RCurl) library(XML) u = "http://humus101.com/?p=2737" a = getURL(u) a # Here - the hebrew is fine. a2 <- htmlParse(a) a2 # Here it is a mess... None of
2012 Sep 14
0
htmlParse pop ups over web pages
Hello All, I am trying to write a routine that loops over some links and parses those links using htmlParse.  The problem is that one of the links may display a pop up window on top of that link's web page.  If there is a pop up, the routine bombs and I get an error message that the url doesn't exist. Does the XML package (or perhaps another package) provide a way to deal with this
2009 Jun 30
1
How to pass parameters to htmlParse Bank of Canada html pages
To get USDCAD rates from Bank of Canada, we first go url <- "http://banqueducanada.ca/en/rates/exchange-avg.html" select 12 months for Rates for the past and click "Get Rates" button. Then the page moves to address <- "http://banqueducanada.ca/cgi-bin/famecgi_fdps" and the rates show in the html page. htmlParse() can read the html document but
2011 Sep 05
2
htmlParse hangs or crashes
Dear colleagues, each time I use htmlParse, R crashes or hangs. The url I'd like to parse is included below as is the results of a series of basic commands that describe what I'm experiencing. The results of sessionInfo() are attached at the bottom of the message. The thing is, htmlTreeParse appears to work just fine, although it doesn't appear to contain the information I need (the
2013 Feb 21
4
Getting htmlParse to work with Hebrew? (on windows)
Hello dear R-help mailing list. Looks like the same issue in Russian: library(RCurl) library(XML) u = " http://www.cian.ru/cat.php?deal_type=2&obl_id=1&room1=1" a = getURL(u) a # Here - the Russian is fine. a2 <- htmlParse(a) a2 # Here it is a mess... None of these seem to fix it: htmlParse(a, encoding = "windows-1251") htmlParse(a, encoding =
2013 Feb 28
0
Scraping data from website---Error in htmlParse: error in creating parser
I'm trying to scrape football projections from accuscore.com for the different positions (right now the projections are set to zeros, but that will change). I can get the QB projections, but I can't get the projections for any of the other positions (e.g., RB). How can I get the RB projections? I'm not sure what the actual website for the RB and other projections is. When I go to
2018 May 23
0
Using R htmlParse() for manipulating URLs to access multiple pages
I am trying to scrape a manual from web. For privacy reasons, I cannot write here the exact URL, anyway, the structure is as follows: https://home.lala.com/bibi/blabla/chapter_i_organization/101_contracts/whatever/,DanaInfo=intranet.lala.com+ https://home.lala.com/bibi/blabla/chapter_i_organization/125_bills/,DanaInfo=intranet.lala.com+
2003 Nov 04
2
4-STABLE b0rked in share/locale/zh_CN.GBK
Murray, Your commits earlier this evening to zh_CN.GB18030 fixed that -STABLE breakage, but zh_CN.GBK appears still to be missing, which causes 'make installworld' to fail. Can you please fix this as well? install -m 644 -o root -g wheel uk_UA.KOI8-U.out /usr/share/locale/uk_UA.KOI8-U/LC_CTYPE install -m 644 -o root -g wheel zh_CN.eucCN.out /usr/share/locale/zh_CN.eucCN/LC_CTYPE
2012 May 19
1
Try Giving Invalid Argument Type Error
Dear R Helpers, I am getting an error message from the try function that I don't understand so I am hoping that someone can help. I am scraping from web pages, but sometimes they disappear. When that happens I need to control for it with some sort of function. This web page is parsed without a problem. exh<-"NASDAQ" tic<-"EGHT"
2012 Oct 17
0
postForm() in RCurl and library RHTMLForms
Hi R Users, I want to get the data from the url given from 10/09/2012 to 15/10/2012. I don't know how to pass the parameters . ....................................................................................................................................... library(RHTMLForms) > > ff = getHTMLFormDescription("
2011 Aug 29
1
reading tables from multiple HTML pages
Hi, beginner to R and was having some problems scraping data from tables in html using the XML package. I have included some code below. I am trying to loop through a series of html pages, each of which contains a single table from which I want to scrape data. However, some of the pages are blank - and so it throws me an error message when it gets to htmlParse(). The loop then closes out and I
2012 Jun 08
0
XML htmlTreeParse fails with no obvious error
Hi all, Sorry for the rather uninformative subject, but the error I get is not very informative either. When using the XML and RCurl package to retrieve the content of an html page, htmlTreeParse fails, printing out the beginning of the HTML: Error in htmlTreeParse(getURL(url)) : File <!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Transitional//EN"
2010 Mar 11
1
parse an HTML page with verbose error message (using XML)
I'm using the function htmlParse() in the XML package, and I need a little bit help on error handling while parsing an HTML page. So far I can use either the default way: # error = xmlErrorCumulator(), by default library(XML) doc = htmlParse("http://www.public.iastate.edu/~pdixon/stat500/") # the error message is: # htmlParseStartTag: invalid element name or the tryCatch()
2015 Oct 05
1
Unable to install Chinese encoding package GBK
Hello, I found it in CentOS 7, the "yum install "Chinese support" is not available. And I need the Chinese simplified encoding with GBK support. Could please suggest how should I install this package in Centos 7? Thanks a lot. Regards, Eric
2011 Apr 21
0
automatic font selection, please help
Hi there, With the helps from this list, I can set specific CJK fonts for character string using text() function. for example: song <- CIDFont("SimSun", "GBK-EUC-H", "GBK", "") postscriptFonts(song = song) postscript("test.ps", height = 7, width =7, family = "Times", fonts = c("song"), horizontal = FALSE, onefile =
2010 Jun 10
0
strange problem:no such file to load -- nokogiri on rails3
I am test the script named test.rb by command "ruby test.rb" and it pass. Then i converte it as rake task. The rake task pass on rails 2.3.4 but fail on rails 3 beta4 with error msg: "no such file to load -- nokogiri" ============================ env :rails 3 beta4, ruby 1.8.7, ubuntu ============================ test.rb ============================ require
2011 Apr 21
2
Chinese segmentation
hello, I have finished reading the papers, and i think it is time to design my project. First step will be determine the input characters are Chinese. i see the past post that cjk-tokenizer is just dealing with UTF-8 and unicode, but i see some other code system such as gbk and big5. i am wondering that should i just deal with UTF-8 and unicode?
2007 Jul 20
2
RSYNC Hung
Hi all, RSYNC Server: CENT OS 5 RSYNC Version: 2.6.8, Running RSYNC Daemon. Client: Windows 2003 Standard, Delta Copy running rsync.exe version 2.6.6 We are backing up 2 files at each site at a total of 3 sites. 1 site is running perfectly, whilst the other 2 are causing constant grief. The backups at these 2 sites are hanging during the RSYNC process at random places. One of the two