Displaying 20 results from an estimated 1300 matches similar to: "[LLVMdev] Question on LLVM based query execution"
2012 Feb 03
0
[LLVMdev] Question on LLVM based query execution
Hi Chang,
> I am developing a Hadoop native runtime, it has C++ APIs and libraries,
> what I want to do is to compile Hive's logical query plan directly to LLVM
> IR or translate Hive's physical query plan to LLVM IR, then run on the
> Hadoop native runtime. As far as I know, Google's tenzing does similar
> things, and a few research papers mention this technique, but
2012 Jan 19
3
[LLVMdev] register allocation
Hi,
My target has special requirements during register allocation - there is both a need to handle register pairing and to never spill a flag result reg-class (which might happen at -O0 for no obvious reason).
Since neither of these issues seems to be supported, I have tried to pre-allocate these registers in the preRA pass. This has resulted in "using undefined physical register"
2009 Jul 31
1
Using R with Hadoop/Hive for Big Data
Hive <http://hadoop.apache.org/hive/> is a data warehouse infrastructure
built on top of Hadoop that provides tools to enable easy data
summarization, adhoc querying and analysis of large datasets data stored in
Hadoop files. It provides a mechanism to put structure on this data and it
also provides a simple query language called QL which is based on SQL and
which enables users familiar with
2006 Nov 22
1
sshd startup error on SCO SR6
Openssh 4.4p1 and 4.5p1
SCO Openserver 6 w/mp1, mp2
While installing v4.5p1, I noticed this sshd syslog error which began
appearing when I upgraded from v4.3p2 to v4.4p1:
Sep 1 17:54:41 tenzing sshd[12837]: error: Bind to port 22 on :: failed: \
Network is unreachable.
sshd -d reports:
debug1: Bind to port 22 on ::.
Bind to port 22 on :: failed: Network is unreachable.
debug1: Bind
2013 May 20
1
Glusterfs-Hadoop
Hi,
Where can I find glusterfs-hadoop-0.20.2-0.1.x86_64.rpm?
The following link is from the Gluster FS Admin Guide, but it doesn't exist:
http://download.gluster.com/pub/gluster/glusterfs/qa-releases/3.3-beta-2/glusterfs-hadoop-0.20.2-0.1.x86_64.rpm
Thanks!
-------------- next part --------------
An HTML attachment was scrubbed...
URL:
2009 May 07
4
problem with conditionals
I''m new to puppet. I''m trying to use some real case examples to better
understand how Puppet works.
Here''s my case:
exec { "usermod -d /home/hadoop -s /bin/bash hadoop":
unless => "test `grep ^hadoop /etc/passwd | awk -F: ''{print
$6}''` == ''/home/hadoop''"
}
The idea is the usermod would only
2013 Mar 11
4
Understanding lustre setup ..
Hello,
I have been reading
http://wiki.lustre.org/images/1/1b/Hadoop_wp_v0.4.2.pdf for setting up
Hadoop over lustre.
Generally in hadoop setup, we have 1 Namenode and various number of datanodes.
If I want to setup the same keeping Lustre as backend, in the document
it is mentioned that:
".............Our experiments run on cluster with 8 nodes in total,
one is mds/namenode, the rest are
2013 Oct 09
2
Error while running MR using rmr2
Hi,
I have trying to run a simple MR program using rmr2 in a single node Hadoop
cluster. Here is the environment for the setup
Ubuntu 12.04 (32 bit)
R (Ubuntu comes with 2.14.1, so updated to 3.0.2)
Installed the latest rmr2 and rhdfs from
here<https://github.com/RevolutionAnalytics/RHadoop/wiki/Downloads>and
the corresponding dependencies
Hadoop 1.2.1
Now I am trying to run a simple MR
2011 Oct 19
1
gluster map/reduce performance..
Hi, all,
i try to check the performance of Map/Reduce of Gluster File system.
Mapper side speed is quite good and it is sometimes faster than hadoop's map job.
But in the Reduce Side job is much slower than hadoop.
i analyze the result and i found the primary reason of slow speed is bad performance in Merging stage.
Would you have any suggestion for this issue
FYI check the blog
2011 Jan 04
5
Allowing puppet to drop privileges for a manifest
Greetings,
Our environment consists of about 600 Redhat Enterprise Linux 3, 4, 5,
and soon 6 servers. We use cfengine 2 currently, but plan on
migrating to puppet. Right now, we have our root-owned cfengine
client running every 15 minutes from cron contacting a single cfservd
server. Additionally, our employees start their own cfengine and
puppet instances on on some servers running under
2013 Nov 20
2
How come that module is not executed in Windows?
I have the following in vagrantfile in WIndows system.
config.vm.provision :puppet do |puppet|
puppet.manifests_path = "manifests"
puppet.manifest_file = "base-hadoop.pp"
puppet.module_path = "modules"
end
when i run vagrant provision, i do see manifest and module folders are
mounted and ssh into vm, I can find files in the following path
2019 Nov 21
2
How to make xapian run in hadoop
Hi all,
We use xapian as the backend of our system. Now the data need be indexed ever-increasing, and the local mode is hard to maintain, so we plan to move the index builder to hadoop. We try to make xapian can be run in hadoop, and now met a problem that there are many seek operations when xapian writes the index files, but the method seek() in hadoop c api only support read, and we blocked by
2010 Oct 08
2
New user - Issue using Generic::Mkuser in the ghoneycutt/generic module.
I''m trying to automatically create users as a requirement for ssh keys
to work. Here is my issue. I am getting this error from the agent. The
SSH part works fine, but it will not create the user due to a
dependency issue. I do not know how to debug this.
err: Could not run Puppet configuration client: Could not find
dependency Generic::Mkuser[hadoop] for Ssh::Authorized_keys[hadoop]
at
2015 Dec 11
2
SVM hadoop
Hola Mª Luz,
Te cuento un poco mi visión:
Lo primero de todo es tener claro qué quiero hacer exactamente en paralelo,
se me ocurren 3 escenarios:
(1) Aplicar un modelo en este caso SVM sobre unos datos muy grandes y por
eso necesito hadoop/spark
(2) Realizar muchos modelos SVM sobre datos pequeños (por ejemplo uno por
usuario) y por eso necesito hadoop/spark para parelilizar estos procesos
2015 Dec 10
3
SVM hadoop
Estimados
Un día leí algo en el siguiente hipervínculo, pero nunca lo use.
http://blog.revolutionanalytics.com/2015/06/using-hadoop-with-r-it-depends.html
Javier Rubén Marcuzzi
De: Carlos J. Gil Bellosta
Enviado: miércoles, 9 de diciembre de 2015 14:33
Para: MªLuz Morales
CC: r-help-es
Asunto: Re: [R-es] SVM hadoop
No, no correrán en paralelo si usas los SVM de paquetes como e1071.
No
2015 Dec 10
2
SVM hadoop
Hola,
Puedes poner un RStudio en Amazon, poner "caret" y a correr....
No sé si tendrás suficiente con lo que te pueda ofrecer Amazon para tu
problema... creo que sí... ;-)....
O directamente hacerlo aquí, que toda esta instalación ya la tienen hecha:
http://www.teraproc.com/front-page-posts/r-on-demand/
Gracias,
Carlos.
El 10 de diciembre de 2015, 14:43, MªLuz Morales <mlzmrls
2008 Aug 21
2
Large data sets with R (binding to hadoop available?)
Dear R community,
I find R fantastic and use R whenever I can for my data analytic
needs. Certain data sets, however, are so large that other tools
seem to be needed to pre-process data such that it can be brought
into R for further analysis.
Questions I have for the many expert contributors on this list are:
1. How do others handle situations of large data sets (gigabytes,
terabytes)
2009 Nov 06
4
Hadoop Cluster on Xen
Hi all,
Has anyone created a Xen cluster to run a hadoop vm cluster?
I would be interested in how it performs
Thanks
Lance
--
This message has been scanned for viruses and
dangerous content by MailScanner, and is
believed to be clean.
_______________________________________________
Xen-users mailing list
Xen-users@lists.xensource.com
http://lists.xensource.com/xen-users
2015 Dec 09
2
SVM hadoop
Buenos días,
alguien sabe si hay alguna manera de implementar una máquina de soporte
vectorial (svm) con R-hadoop??
Mi interés es hacer procesamiento big data con svm. Se que en R, existen
los paquetes {RtextTools} y {e1071} que permiten hacer svm. Pero no estoy
segura de que el algoritmo sea paralelizable, es decir, que pueda correr en
paralelo a través de la plataforma R-hadoop.
Muchas
2016 Jun 15
5
Hadoop
Hola buenas,
me preguntaba si alguno usa hadoop Spark en su día día y si me podíais
recomendar un buen curso para empezar. Estuve en la charla de meetup de
madrid hace unos meses de Rspark y estuvo bien, ahora me preguntaba si es
posible profundizar.
Pero me gustaría tener recomendaciones de cualquier material que podáis
recomendar, cursos de coursera que hayais hecho, libros que hayais leido,