similar to: s3 as mysql directory

Displaying 20 results from an estimated 9000 matches similar to: "s3 as mysql directory"

2011 Jul 31
4
cron jobs not running
hello list!! I'm attempting to find out why this cron job isn't running. the host is centos 5.6 on i386 just so you know. 0 3 * * * /bin/alldb > /home/bluethundr/backupdb/alldb-$(date +%Y%m%d%H%S).sql The job is meant to backup all the databases on a mysql server every day at 3 am. I have checked the backup directory and nothing is showing up for those
2012 Mar 05
5
cron job not running
hello list, I am attempting to backup a centos 5.4 (x86_64) server running mysql with a cron job. Here's how the cron job looks: [root at cloud:/home/bluethundr/backupdb] #crontab -l * 3 * * * /usr/bin/mysqldump jfwiki > /home/bluethundr/backupdb/wiki-$(date +%Y%m%d).sql However if I run the command from the command line it seems to work fine. If I grep syslog for cron this is what
2015 Aug 15
2
wordpess can't connect to DB but mediawiki can
> > "Use" that db and then issue: > select * from db where Db='jfwiki' or Db='jokefire' order by Host; Well yeah. I used the mysql database before I issued that command. MariaDB [(none)]> use mysql Reading table information for completion of table and column names You can turn off this feature to get a quicker startup with -A Database changed Then if I
2010 Apr 12
1
Amazon S3FS Automounter of sorts
Hi, I?ve just started using Amazon S3 for storage and have used s3fs to mount the buckets on the file system. 1. I was wondering if anyone has a better method ? even though this meets my current needs. 2. if there is an automounter that could be tweaked to work with s3fs. The entries in my fstab are rather different to the normal disk based ones, s3fs#BUCKETNAME /mnt/s3/BUCKETNAME fuse
2008 Feb 14
1
Backing up remote system
I have an Overland Arcvault tape library, a CentOS 5 box, a Windows XP system, and RAID box that supports NFS and CIFS. The RAID box is remotely located and acts as central file storage. I might normally use dump to perform backups, but as was learned here, and on dump's man page, dump doesn't support remote file systems such as NFS or CIFS. So, for now, I've connected the
2009 Oct 19
4
Backing up a centos serveR? how to ? Newbie question..
Hi all, excuse my newbie question but how can i backup my centos server? i have a dozzen of virtual hosts over it as well as substantial database entries.. i've backed up the following directories using rsync: workspace/ /etc/httpd/ /etc/apache-tomcat-6.0.20/ /usr/lib/mysql/ /var/lib/mysql though i'm still not at ease with this, so i decided to do a some sort of clone or something
2015 Aug 15
2
wordpess can't connect to DB but mediawiki can
> > [this isn't really a centos issue, even if you're using centos, > which isn't obvious. that said ...] Yeah that's true. But this list tends to be rather helpful for general problems that are less specific to centos. Sometimes. :) Really seems to depend... Incidentally I am using centos on all hosts: #cat /etc/redhat-release CentOS Linux release 7.0.1406 (Core) OK
2010 Aug 08
2
paperclip save to disk and s3
I have a standard Paperclip setup that saves a file to my disk. In addition I would also like the file saved to my amazon s3 bucket. [code] after_save :copy_to_s3 def copy_to_s3 has_attached_file :photo, :storage => :s3, :s3_credentials => "#{RAILS_ROOT}/config/s3.yml", :styles => { :thumb => "100x100#", :small => "750x750>"
2015 Jun 21
2
puppet files denied by SELinux
Hi all, Thanks for all your suggestions. Here's where I'm at with this. Can you give details about your puppetmasterd setup ? it seems that > you're using Foreman as puppet ENC. > Yes, I'm on foreman 1.7.4 and puppet 3.75. You are correct that I'm using foreman, sorry I hadn't thought to mention it! > Foreman works fine with selinux enabled : that's what
2015 Aug 16
2
wordpess can't connect to DB but mediawiki can
> > You were doing this (looking at the mysql.db table) on your > "db.example.com" machine, correct? db.example.com is a load balanced VIP. The VIP is being handled by keepalived and HA/Proxy. There are two DB's setup in master/master replication. The two databases and two load balancers are on AWS. The web server and varnish servers are on digital ocean. I setup a grant
2015 Jun 20
2
puppet files denied by SELinux
Hey folks, Ok so I'm having another issue with SELinux. However I think I'm pretty close to a solution and just need a nudge in the right directtion. I wrote a puppet module that gets systems into bacula backups. Part of the formula is to distribute key/cert pairs with permissions that allow bacula to read them so that bacula can talk to the host over TLS. It's pretty slick, I must
2010 Oct 29
8
Amazon s3 - paperclip - Rails3 - basic setup Noob simple problems
Hi all, I''ve got a feeling from other posts that this should be very simple, but believe me I''ve looked around many tutorials / blogs, but still cant even get the most basic setup working with paperclip and s3. I am however pretty new to Rails, so could be doing something dumb. ------- Firstly, to install and use paperclip and aws-s3 I''ve got them both in by Gemfile:
2011 Jul 09
2
Zip files on Amazon s3 bucket
Ive recently moved my files to amazon S3 storage. There is some functionality to allow an archive of files to be downloaded to a user via delayed job. So basically a project may have 10 documents in 10 different folders in the s3 bucket. What i need to do is zip the files up into one zip file and then send the user a link to download them. Im using https://github.com/qoobaa/s3 and paperclip to
2013 Jan 25
4
store pdf directly to amazon
Hello Everyone, I am using wicked pdf and rails 3.2.8. I can generate pdf (html to pdf). How can i store my pdf file into amazon s3 without storing in *Local.* * * *Please help* -- You received this message because you are subscribed to the Google Groups "Ruby on Rails: Talk" group. To unsubscribe from this group and stop receiving emails from it, send an email to
2012 Mar 24
5
your advice on backup procedure
Hello all, I am down to my last hurdle of my project, backups. I am thinking of three different ways to go and wanted to ask for input on what you think is the better choice. Not asking for 'how to' but more of 'what is best in your experience' The scenario... centos server acting as a virtual host. Virtual machines are webservers and dns servers. All on one machine, all running
2010 Aug 14
6
best ways to do mysql backup
we have multiple servers approx 10 and each has about 100 GB of data in the /var/lib/mysql dir , excluding tar , mysqldump and replication how do we take backup for these databases on to a remote machine and store them datewise , ( the remote machine is a 2TB HDD ) currently tar is not feasible as the data is too huge and the same goes with mysqldump suggestion will be of great help --
2013 Mar 12
8
Heroku db id Starting with 1000
Hello All, How do I change the heroku database( already existed ) table id to start from say 1000? I tried with a fresh database on local & it works with this:- *execute "ALTER users orders AUTO_INCREMENT = 1000"* Thanks, Avinash -- You received this message because you are subscribed to the Google Groups "Ruby on Rails: Talk" group. To unsubscribe from
2019 Nov 04
3
Resumen de R-help-es, Vol 129, Envío 1
Hola, Alguien con experiencia en cargar ficheros csv en aws SageMaker con Instancia Notebook? Mabel Garc?a [[alternative HTML version deleted]]
2012 Jul 03
4
s3fs vs. zfs
Hi all.. Is possible use the s3fs with ZFS (freebsd) ? how ? Thanks -- Luiz Gustavo Costa (Powered by BSD) *+*+*+*+*+*+*+*+*+*+*+*+*+*+*+*+*+*+*+*+*+*+*+*+*+ mundoUnix - Consultoria em Software Livre http://www.mundounix.com.br ICQ: 2890831 / MSN: contato at mundounix.com.br Tel: 55 (21) 4063-7110 / 8194-1905 / (11) 4063-0407 Blog: http://www.luizgustavo.pro.br
2014 Nov 10
3
backup incrementals on mysql
Hi all. I usally make backups of databases mysql. I make buckups of all datbase for example: mysqldump -u user -ppassword name_db > backups.sql also I make backups just its schema for example mysqldump -u user -ppassword name_db --no-data > backups.sql but now I need a backups incrementals, because the size of DB is very big (500 mb) How to make this? regards -- *Atte.