similar to: Samba and CEPH

Displaying 20 results from an estimated 3000 matches similar to: "Samba and CEPH"

2016 Jan 08
1
Samba & Ceph
On 2016-01-08 at 09:31 -0800, Jeremy Allison wrote: > On Fri, Jan 08, 2016 at 04:26:24PM +0100, Dirk Laurenz wrote: > > Hello List, > > > > as anyone tried to install samba with/ontop on a ceph cluster? > > Try compiling and setting up with vfs_ceph. Correct, that's basically it. > Needs some more work, but should work. Some posix features are not quite there
2016 Jan 08
2
Samba & Ceph
Hello List, as anyone tried to install samba with/ontop on a ceph cluster? Regards, Dirk
2018 May 23
3
ceph_vms performance
Hi, I'm testing out ceph_vms vs a cephfs mount with a cifs export. I currently have 3 active ceph mds servers to maximise throughput and when I have configured a cephfs mount with a cifs export, I'm getting a reasonable benchmark results. However, when I tried some benchmarking with the ceph_vms module, I only got a 3rd of the comparable write throughput. I'm just wondering if
2019 Nov 07
2
samba performance when writing lots of small files
hi jeremy / all, On 11/6/19 10:39 PM, Jeremy Allison wrote: > This is re-exporting via ceph whilst creating 1000 files, > yes ? What timings do you get when doing this via Samba > onto a local ext4/xfs/btrfs/zfs filesystem ? yes, creating 10k small files. doing the same on a local ssd, formatted with an ext4 fs without any special options: root at plattentest:/mnt-ssd/os# time for s in
2020 Sep 21
2
ceph vfs can't find specific path
Hello Using two file server with samba 4.12.6 running as a CTDB cluster and trying to share a specific path on a cephfs. After loading the config the ctdb log shows the following error: ctdb-eventd[248]: 50.samba: ERROR: samba directory "/plm" not available Here is my samba configuration: [global] clustering = Yes netbios name = FSCLUSTER realm = INT.EXAMPLE.COM registry
2018 May 16
1
[ceph-users] dovecot + cephfs - sdbox vs mdbox
Thanks Jack. That's good to know. It is definitely something to consider. In a distributed storage scenario we might build a dedicated pool for that and tune the pool as more capacity or performance is needed. Regards, Webert Lima DevOps Engineer at MAV Tecnologia *Belo Horizonte - Brasil* *IRC NICK - WebertRLZ* On Wed, May 16, 2018 at 4:45 PM Jack <ceph at jack.fr.eu.org> wrote:
2018 Oct 08
3
vfs_ceph quota support?
Hi Folks, is the vfs_ceph supporting quota set on a directory inside cephfs? Regards Felix -- Forschungszentrum Jülich GmbH 52425 Jülich Sitz der Gesellschaft: Jülich Eingetragen im Handelsregister des Amtsgerichts Düren Nr. HR B 3498 Vorsitzender des Aufsichtsrats: MinDir. Dr. Karl Eugen Huthmacher Geschäftsführung: Prof. Dr.-Ing. Wolfgang Marquardt (Vorsitzender), Karsten Beneke (stellv.
2014 Jan 15
2
Ceph RBD locking for libvirt-managed LXC (someday) live migrations
Hi, I'm trying to build an active/active virtualization cluster using a Ceph RBD as backing for each libvirt-managed LXC. I know live migration for LXC isn't yet possible, but I'd like to build my infrastructure as if it were. That is, I would like to be sure proper locking is in place for live migrations to someday take place. In other words, I'm building things as if I were
2018 May 16
2
dovecot + cephfs - sdbox vs mdbox
I'm sending this message to both dovecot and ceph-users ML so please don't mind if something seems too obvious for you. Hi, I have a question for both dovecot and ceph lists and below I'll explain what's going on. Regarding dbox format (https://wiki2.dovecot.org/MailboxFormat/dbox), when using sdbox, a new file is stored for each email message. When using mdbox, multiple
2015 Mar 31
2
couple of ceph/rbd questions
Hi, I've recently been working on setting up a set of libvirt compute nodes that will be using a ceph rbd pool for storing vm disk image files. I've got a couple of issues I've run into. First, per the standard ceph documentation examples [1], the way to add a disk is to create a block in the VM definition XML that looks something like this: <disk type='network'
2023 Dec 14
2
Gluster -> Ceph
Hi all, I am looking in to ceph and cephfs and in my head I am comparing with gluster. The way I have been running gluster over the years is either a replicated or replicated-distributed clusters. The small setup we have had has been a replicated cluster with one arbiter and two fileservers. These fileservers has been configured with RAID6 and that raid has been used as the brick. If disaster
2003 Nov 09
1
Solved: Samba 3 w/ADS on Slackware
Hi folks: We have finally solved the problem. We have figured out how to correctly install Samba 3 with Active Directory Support on Slackware (and presumably on any other non-RPM based distro). The howto has been published on my website. Please visit http://www.rongage.org/manual_samba_howto.html for the step-by-step instructions on how to manually install Samba 3 from scratch with Active
2017 Sep 22
3
librmb: Mail storage on RADOS with Dovecot
Hi ceph-ers, The email below was posted on the ceph mailinglist yesterday by Wido den Hollander. I guess this could be interesting for user here as well. MJ -------- Forwarded Message -------- Subject: [ceph-users] librmb: Mail storage on RADOS with Dovecot Date: Thu, 21 Sep 2017 10:40:03 +0200 (CEST) From: Wido den Hollander <wido at 42on.com> To: ceph-users at ceph.com Hi, A tracker
2023 Dec 14
2
Gluster -> Ceph
Big raid isn't great as bricks. If the array does fail, the larger brick means much longer heal times. My main question I ask when evaluating storage solutions is, "what happens when it fails?" With ceph, if the placement database is corrupted, all your data is lost (happened to my employer, once, losing 5PB of customer data). With Gluster, it's just files on disks, easily
2003 Mar 08
7
IAX on windows
I know this has come up before but... Has anyone done anything to get an IAX client built on Windows? I thought someone had started one, but I haven't heard anything about it since - and that was months ago? Anyone have any idea what the status is? -- Ron Gage - Saginaw, Michigan I am looking for work - resume at http://www.rongage.org/resume.doc Electrical Engineering, Linux Programming,
2003 Oct 26
1
SAMBA 3 and ADS still doesn't work
Been working on this for the past 3 weeks and it STILL doesn't work. Machine 1: Samba 3 on Slackware 9 - member server Machine 2: Windows 2000 Advanced Server - PDC, AD Machine 1 has successfully joined the domain without incident - it shows up in AD-Users & Computers. kinit ADMINISTRATOR@MACHINE2 works on machine 1, accepts the correct password, and returns no errors. Machine 2 sees
2018 May 16
0
[ceph-users] dovecot + cephfs - sdbox vs mdbox
Hi, some time back we had similar discussions when we, as an email provider, discussed to move away from traditional NAS/NFS storage to Ceph. The problem with POSIX file systems and dovecot is that e.g. with mdbox only around ~20% of the IO operations are READ/WRITE, the rest are metadata IOs. You will not change this with using CephFS since it will basically behave the same way as e.g. NFS. We
2015 Jan 26
1
how to add new vfs module to build...
On 1/24/2015 3:03 PM, Volker Lendecke wrote: > On Sat, Jan 24, 2015 at 01:35:40PM -0500, Ed Sutter wrote: >> Hi, >> I'm working with version 4.1.6 to create a new VFS module. >> I've been experimenting by just modifying one of the existing >> module source files (vfs_full_audit.c) and that rebuilds as expected. >> >> I want to step back now and create
2023 Jun 12
2
virsh not connecting to libvertd ?
Just found my issue. After I removed the cephfs mounts it worked! I will debug ceph. I assumed because I could touch files on mounted cephfs it was working. Now virsh list works! thanks jerry Lars Kellogg-Stedman > On Tue, Jun 06, 2023 at 04:56:38PM -0400, Jerry Buburuz wrote: >> Recently both virsh stopped talking to the libvirtd. Both stopped within >> a >> few days of
2016 May 30
1
Re: migrate local storage to ceph | exchanging the storage system
On 05/30/2016 09:07 AM, Dominique Ramaekers wrote: >> root@host_a:~# virsh migrate --verbose --p2p --copy-storage-all --persistent -- >> change-protection --abort-on-error --undefinesource --live domain >> qemu+ssh://root@host_b/system --xml domain.ceph.xml > > Weird: The domain should be persistent Well, the domain is persistent. But the changes i did to domain.ceph.xml