Displaying 20 results from an estimated 157 matches for "hadoops".
Did you mean:
hadoop
2013 May 20
1
Glusterfs-Hadoop
Hi,
Where can I find glusterfs-hadoop-0.20.2-0.1.x86_64.rpm?
The following link is from the Gluster FS Admin Guide, but it doesn't exist:
http://download.gluster.com/pub/gluster/glusterfs/qa-releases/3.3-beta-2/glusterfs-hadoop-0.20.2-0.1.x86_64.rpm
Thanks!
-------------- next part --------------
An HTML attachment was scrubbed...
URL:
2009 May 07
4
problem with conditionals
I''m new to puppet. I''m trying to use some real case examples to better
understand how Puppet works.
Here''s my case:
exec { "usermod -d /home/hadoop -s /bin/bash hadoop":
unless => "test `grep ^hadoop /etc/passwd | awk -F: ''{print
$6}''` == ''/home/hadoop''"
}
The idea is the usermod would only
2013 Oct 09
2
Error while running MR using rmr2
Hi,
I have trying to run a simple MR program using rmr2 in a single node Hadoop
cluster. Here is the environment for the setup
Ubuntu 12.04 (32 bit)
R (Ubuntu comes with 2.14.1, so updated to 3.0.2)
Installed the latest rmr2 and rhdfs from
here<https://github.com/RevolutionAnalytics/RHadoop/wiki/Downloads>and
the corresponding dependencies
Hadoop 1.2.1
Now I am trying to run a simple MR
2013 Mar 11
4
Understanding lustre setup ..
Hello,
I have been reading
http://wiki.lustre.org/images/1/1b/Hadoop_wp_v0.4.2.pdf for setting up
Hadoop over lustre.
Generally in hadoop setup, we have 1 Namenode and various number of datanodes.
If I want to setup the same keeping Lustre as backend, in the document
it is mentioned that:
".............Our experiments run on cluster with 8 nodes in total,
one is mds/namenode, the rest are
2019 Nov 21
2
How to make xapian run in hadoop
Hi all,
We use xapian as the backend of our system. Now the data need be indexed ever-increasing, and the local mode is hard to maintain, so we plan to move the index builder to hadoop. We try to make xapian can be run in hadoop, and now met a problem that there are many seek operations when xapian writes the index files, but the method seek() in hadoop c api only support read, and we blocked by
2008 Aug 21
2
Large data sets with R (binding to hadoop available?)
Dear R community,
I find R fantastic and use R whenever I can for my data analytic
needs. Certain data sets, however, are so large that other tools
seem to be needed to pre-process data such that it can be brought
into R for further analysis.
Questions I have for the many expert contributors on this list are:
1. How do others handle situations of large data sets (gigabytes,
terabytes)
2011 Oct 19
1
gluster map/reduce performance..
Hi, all,
i try to check the performance of Map/Reduce of Gluster File system.
Mapper side speed is quite good and it is sometimes faster than hadoop's map job.
But in the Reduce Side job is much slower than hadoop.
i analyze the result and i found the primary reason of slow speed is bad performance in Merging stage.
Would you have any suggestion for this issue
FYI check the blog
2009 Jul 31
1
Using R with Hadoop/Hive for Big Data
Hive <http://hadoop.apache.org/hive/> is a data warehouse infrastructure
built on top of Hadoop that provides tools to enable easy data
summarization, adhoc querying and analysis of large datasets data stored in
Hadoop files. It provides a mechanism to put structure on this data and it
also provides a simple query language called QL which is based on SQL and
which enables users familiar with
2013 Nov 20
2
How come that module is not executed in Windows?
I have the following in vagrantfile in WIndows system.
config.vm.provision :puppet do |puppet|
puppet.manifests_path = "manifests"
puppet.manifest_file = "base-hadoop.pp"
puppet.module_path = "modules"
end
when i run vagrant provision, i do see manifest and module folders are
mounted and ssh into vm, I can find files in the following path
2010 Oct 08
2
New user - Issue using Generic::Mkuser in the ghoneycutt/generic module.
I''m trying to automatically create users as a requirement for ssh keys
to work. Here is my issue. I am getting this error from the agent. The
SSH part works fine, but it will not create the user due to a
dependency issue. I do not know how to debug this.
err: Could not run Puppet configuration client: Could not find
dependency Generic::Mkuser[hadoop] for Ssh::Authorized_keys[hadoop]
at
2011 Jan 04
5
Allowing puppet to drop privileges for a manifest
Greetings,
Our environment consists of about 600 Redhat Enterprise Linux 3, 4, 5,
and soon 6 servers. We use cfengine 2 currently, but plan on
migrating to puppet. Right now, we have our root-owned cfengine
client running every 15 minutes from cron contacting a single cfservd
server. Additionally, our employees start their own cfengine and
puppet instances on on some servers running under
2015 Dec 11
2
SVM hadoop
Hola Mª Luz,
Te cuento un poco mi visión:
Lo primero de todo es tener claro qué quiero hacer exactamente en paralelo,
se me ocurren 3 escenarios:
(1) Aplicar un modelo en este caso SVM sobre unos datos muy grandes y por
eso necesito hadoop/spark
(2) Realizar muchos modelos SVM sobre datos pequeños (por ejemplo uno por
usuario) y por eso necesito hadoop/spark para parelilizar estos procesos
2015 Dec 10
3
SVM hadoop
Estimados
Un día leí algo en el siguiente hipervínculo, pero nunca lo use.
http://blog.revolutionanalytics.com/2015/06/using-hadoop-with-r-it-depends.html
Javier Rubén Marcuzzi
De: Carlos J. Gil Bellosta
Enviado: miércoles, 9 de diciembre de 2015 14:33
Para: MªLuz Morales
CC: r-help-es
Asunto: Re: [R-es] SVM hadoop
No, no correrán en paralelo si usas los SVM de paquetes como e1071.
No
2015 Dec 10
2
SVM hadoop
Hola,
Puedes poner un RStudio en Amazon, poner "caret" y a correr....
No sé si tendrás suficiente con lo que te pueda ofrecer Amazon para tu
problema... creo que sí... ;-)....
O directamente hacerlo aquí, que toda esta instalación ya la tienen hecha:
http://www.teraproc.com/front-page-posts/r-on-demand/
Gracias,
Carlos.
El 10 de diciembre de 2015, 14:43, MªLuz Morales <mlzmrls
2013 Sep 23
0
Unable to execute Java MapReduce (Hadoop) code from R using rJava
Hi All,
I have written a Java MapReduce code that runs on Hadoop. My intention is
to create an R package which will call the Java code and execute the job.
Hence, I have written a similar R function. But when I call this function
from R terminal, the Hadoop job is not running. Its just printing few lines
of warning messages and does nothing further. Here is the execution
scenario:
*>
2015 Dec 09
2
SVM hadoop
Buenos días,
alguien sabe si hay alguna manera de implementar una máquina de soporte
vectorial (svm) con R-hadoop??
Mi interés es hacer procesamiento big data con svm. Se que en R, existen
los paquetes {RtextTools} y {e1071} que permiten hacer svm. Pero no estoy
segura de que el algoritmo sea paralelizable, es decir, que pueda correr en
paralelo a través de la plataforma R-hadoop.
Muchas
2012 Nov 07
2
R + Hadoop on Amazon
Hello All,
Having some issue with local machine, I need to locate myself on Amazon
for running R and Hadoop with Amazon instance. After searching a lot, I
can't able to take a decision for choosing Image for Amazon instance. Can any
one using R + Hadoop on Amazon.
Thanks
[[alternative HTML version deleted]]
2009 Nov 06
4
Hadoop Cluster on Xen
Hi all,
Has anyone created a Xen cluster to run a hadoop vm cluster?
I would be interested in how it performs
Thanks
Lance
--
This message has been scanned for viruses and
dangerous content by MailScanner, and is
believed to be clean.
_______________________________________________
Xen-users mailing list
Xen-users@lists.xensource.com
http://lists.xensource.com/xen-users
2019 Nov 22
0
How to make xapian run in hadoop
On Thu, Nov 21, 2019 at 10:20:19AM +0800, ??? wrote:
> We use xapian as the backend of our system. Now the data need be
> indexed ever-increasing, and the local mode is hard to maintain, so we
> plan to move the index builder to hadoop. We try to make xapian can be
> run in hadoop, and now met a problem that there are many seek
> operations when xapian writes the index files, but
2010 Dec 24
1
Running scripts in hadoop
R-help group,
I'm looking for some assistance on using an R-script to read STDIN from
hadoop.
Example, say I have two tables. One is a student table, the other is a class
roster table (tables join on student_id). Student SAT score is in the
student table, whether the student passed or not is in the roster table.
So to determine if a student passed or failed based on their SAT score, I'd