similar to: running icecast on top of hadoop

Displaying 20 results from an estimated 2000 matches similar to: "running icecast on top of hadoop"

2011 Dec 09
1
Samba 3.0, fuse-hdfs and write problems
Hi folks, I research at the moment a connection over fuse and samba for hadoop-cluster. Its my private playground, but I have some issues I can't figure out why they happen. I use RHEL5.7, packetlist: rpm -qa|grep samba samba-3.0.33-3.29.el5_7.4.x86_64 samba-common-3.0.33-3.29.el5_7.4.x86_64 2 servers, one provides samba-shares and one I use as client. Samba-Server (HOST2): FuseFS:
2016 Apr 22
1
Storage cluster advise, anybody?
Hi Valeri On Fri, Apr 22, 2016 at 10:24 PM, Digimer <lists at alteeve.ca> wrote: > On 22/04/16 03:18 PM, Valeri Galtsev wrote: >> Dear Experts, >> >> I would like to ask everybody: what would you advise to use as a storage >> cluster, or as a distributed filesystem. >> >> I made my own research of what I can do, but I hit a snag with my >>
2009 Nov 06
4
Hadoop Cluster on Xen
Hi all, Has anyone created a Xen cluster to run a hadoop vm cluster? I would be interested in how it performs Thanks Lance -- This message has been scanned for viruses and dangerous content by MailScanner, and is believed to be clean. _______________________________________________ Xen-users mailing list Xen-users@lists.xensource.com http://lists.xensource.com/xen-users
2016 Apr 15
1
help on moving data from local to HDFS using RODBC in R
Hi, I have requirement to move the data from Linux local path( ex /home/user/sample.txt) to hadoop HDFS using RODBC in R I knew that we can move the data using rhive comamnds like *rhive.put* and *rhive.get *but looking for similar commands using RODBC as well. I would appreciate for your inputs. Regards, Divakar Phoenix, USA [[alternative HTML version deleted]]
2013 May 09
0
How to run TestDFSIO on hadoop running lustre?
I generally run TestDFSIO for Hadoop running on HDFS as shown: hadoop jar hadoop-*test*.jar TestDFSIO -write -nrFiles 10 -fileSize 1000 How to run it on lustre filesystem? -------------- next part -------------- An HTML attachment was scrubbed... URL: http://lists.lustre.org/pipermail/lustre-discuss/attachments/20130509/821dc4bb/attachment.html
2013 Sep 23
0
Unable to execute Java MapReduce (Hadoop) code from R using rJava
Hi All, I have written a Java MapReduce code that runs on Hadoop. My intention is to create an R package which will call the Java code and execute the job. Hence, I have written a similar R function. But when I call this function from R terminal, the Hadoop job is not running. Its just printing few lines of warning messages and does nothing further. Here is the execution scenario: *>
2011 Oct 19
1
gluster map/reduce performance..
Hi, all, i try to check the performance of Map/Reduce of Gluster File system. Mapper side speed is quite good and it is sometimes faster than hadoop's map job. But in the Reduce Side job is much slower than hadoop. i analyze the result and i found the primary reason of slow speed is bad performance in Merging stage. Would you have any suggestion for this issue FYI check the blog
2011 Jun 13
0
Hadoop Hive output read into R
All, I am using a pretty crude method to get data out of HDFS via Hive and into R and was curious about alternatives that the group has explored. Basically, I run a system command that runs a hive statement and writes the returned data to a delimited file. Then, I read that file into an object and continue. For example: hive.script <- "select * from orders where date =
2012 Feb 06
1
using file in hdfs for data mining algorithms in r
hi all, i am new to r , i am trying to run data mining algorithms using map reduce framework.. * *i have few basic doubts* *1. can i give file in hdfs to kmeans( ) ? ?I tried as > file1 = hdfs.file("testdata/synthetic_control.data") > isf = hdfs.read(lsf,5242880,0) >l = kmeans(isf,2,10) its not working.. please help 2. How to access the file in hdfs and give as input to
2013 Oct 09
1
How to write R data frame to HDFS using rhdfs?
Hello, I am trying to write the default "OrchardSprays" R data frame into HDFS using the "rhdfs" package. I want to write this data frame directly into HDFS without first storing it into any file in local file system. Which rhdfs command i should use? Can some one help me? I am very new to R and rhdfs. Regards, Gaurav [[alternative HTML version deleted]]
2019 Nov 21
2
How to make xapian run in hadoop
Hi all, We use xapian as the backend of our system. Now the data need be indexed ever-increasing, and the local mode is hard to maintain, so we plan to move the index builder to hadoop. We try to make xapian can be run in hadoop, and now met a problem that there are many seek operations when xapian writes the index files, but the method seek() in hadoop c api only support read, and we blocked by
2013 Mar 11
4
Understanding lustre setup ..
Hello, I have been reading http://wiki.lustre.org/images/1/1b/Hadoop_wp_v0.4.2.pdf for setting up Hadoop over lustre. Generally in hadoop setup, we have 1 Namenode and various number of datanodes. If I want to setup the same keeping Lustre as backend, in the document it is mentioned that: ".............Our experiments run on cluster with 8 nodes in total, one is mds/namenode, the rest are
2009 Jul 31
1
Using R with Hadoop/Hive for Big Data
Hive <http://hadoop.apache.org/hive/> is a data warehouse infrastructure built on top of Hadoop that provides tools to enable easy data summarization, adhoc querying and analysis of large datasets data stored in Hadoop files. It provides a mechanism to put structure on this data and it also provides a simple query language called QL which is based on SQL and which enables users familiar with
2010 Jun 16
6
clustered file system of choice
Hi all, I am just trying to consider my options for storing a large mass of data (tens of terrabytes of files) and one idea is to build a clustered FS of some kind. Has anybody had any experience with that? Any recommendations? Thanks in advance for any and all advice. Boris.
2008 Aug 21
2
Large data sets with R (binding to hadoop available?)
Dear R community, I find R fantastic and use R whenever I can for my data analytic needs. Certain data sets, however, are so large that other tools seem to be needed to pre-process data such that it can be brought into R for further analysis. Questions I have for the many expert contributors on this list are: 1. How do others handle situations of large data sets (gigabytes, terabytes)
2015 Apr 22
2
Lossless stream from MPD...
I am not familular with liquidsoap. Is it something I can use with MPD? On Wed, Apr 22, 2015 at 1:10 PM, Eduardo Martinez <erm13martinez at gmail.com> wrote: > Icecast can definitly handle OggFLAC. I usually use liquidsoap to stream > flac to icecast. > On Apr 22, 2015 12:52 PM, "David Martinez" <david at sojournmusic.net> wrote: > >> Any way to setup a
2013 May 20
1
Glusterfs-Hadoop
Hi, Where can I find glusterfs-hadoop-0.20.2-0.1.x86_64.rpm? The following link is from the Gluster FS Admin Guide, but it doesn't exist: http://download.gluster.com/pub/gluster/glusterfs/qa-releases/3.3-beta-2/glusterfs-hadoop-0.20.2-0.1.x86_64.rpm Thanks! -------------- next part -------------- An HTML attachment was scrubbed... URL:
2015 Apr 22
2
Lossless stream from MPD...
Will this work? # Stream it out output.icecast(%flac, host = "localhost", port = 8000, password = "hackme", mount = "basic-radio.flac", radio) On Apr 22, 2015 1:22 PM, "Eduardo Martinez" <erm13martinez at gmail.com> wrote: > Liquidsoap is another encoder simular to ices or MPD. Liquidsoap doesn't > work directly with MPD. > On Apr 22,
2012 Nov 07
2
R + Hadoop on Amazon
Hello All, Having some issue with local machine, I need to locate myself on Amazon for running R and Hadoop with Amazon instance. After searching a lot, I can't able to take a decision for choosing Image for Amazon instance. Can any one using R + Hadoop on Amazon. Thanks [[alternative HTML version deleted]]
2015 Dec 09
2
SVM hadoop
Buenos días, alguien sabe si hay alguna manera de implementar una máquina de soporte vectorial (svm) con R-hadoop?? Mi interés es hacer procesamiento big data con svm. Se que en R, existen los paquetes {RtextTools} y {e1071} que permiten hacer svm. Pero no estoy segura de que el algoritmo sea paralelizable, es decir, que pueda correr en paralelo a través de la plataforma R-hadoop. Muchas