similar to: ocfs2 problem on ctdb cluster

Displaying 20 results from an estimated 1100 matches similar to: "ocfs2 problem on ctdb cluster"

2019 Jan 29
2
[monorepo] Much improved downstream zipping tool available
Björn Pettersson A <bjorn.a.pettersson at ericsson.com> writes: > In the new monorepo UC1 may or may not be a parent to UL1. > We could actually have something like this: > > UL4->UC2->UL3->UL2->UL1->UL0->UC1 > > Our DL1 commit should preferably have UL1 as parent after > conversion > > UL4->UC2->UL3->UL2->UL1->UL0->UC1 >
2019 Jan 30
2
[monorepo] Much improved downstream zipping tool available
Björn Pettersson A <bjorn.a.pettersson at ericsson.com> writes: > In llvm (split) we have: > > UL4->UL3->UL2->UL1->UL0 > \ > ...->DL2->DL1 > > In clang (split) we have: > > UC4->UC3->UC2->UC1->UC0 > \ > ...->DC2->DC1 > > > DL1 is a commit that updates the
2014 Aug 15
1
smbd crash: 4.1.11
Hi Random crashes. It looks as if it's cups, but we aren't printing, nor have it configured. [global] workgroup = ALTEA realm = ALTEA.SITE security = ADS kerberos method = secrets and keytab netbios name = SMBCLUSTER #disable netbios = Yes clustering = Yes ctdbd socket = /usr/local/var/run/ctdb/ctdbd.socket [users] path = /cluster/users read only = No [profiles] path = /cluster/profiles
2019 Jan 29
2
[monorepo] Much improved downstream zipping tool available
He all, I've updated the downstream fork zipping tool that I posted about last November [1]. It is much improved in every way. The most important enhancements are: - Does a better job of simplifying history - Handles nested submodules - Will put non-submodule-update content in a subdirectory of the monorepo - Updates tags In addition there are plenty of the requisite bug fixes. The
2014 Jul 21
0
CTDB no secrets.tdb created
Hi 2 node ctdb 2.5.3 on Ubuntu 14.04 nodes apparmor teardown and firewall and stopped dead The IP takeover is working fine between the nodes: Jul 21 14:12:03 uc1 ctdbd: recoverd:Trigger takeoverrun Jul 21 14:12:03 uc1 ctdbd: recoverd:Takeover run starting Jul 21 14:12:04 uc1 ctdbd: Takeover of IP 192.168.1.81/24 on interface bond0 Jul 21 14:12:04 uc1 ctdbd: Takeover of IP 192.168.1.80/24 on
2011 Dec 20
1
OCFS2 problems when connectivity lost
Hello, We are having a problem with a 3-node cluster based on Pacemaker/Corosync with 2 primary DRBD+OCFS2 nodes and a quorum node. Nodes run on Debian Squeeze, all packages are from the stable branch except for Corosync (which is from backports for udpu functionality). Each node has a single network card. When the network is up, everything works without any problems, graceful shutdown of
2011 Mar 03
1
OCFS2 1.4 + DRBD + iSCSI problem with DLM
An HTML attachment was scrubbed... URL: http://oss.oracle.com/pipermail/ocfs2-users/attachments/20110303/0fbefee6/attachment.html
2007 Jul 07
2
Adding new nodes to OCFS2?
I looked around, found older post which seems not applicable anymore. I have a cluster of 2 nodes right now, which has 3 OCFS2 file systems. All the file systems were formatted with 4 node slots. I added the two news nodes (by hand, by ocfs2console and o2cb_ctl), so my /etc/ofcfs/cluster.conf looks right: node: ip_port = 7777 ip_address = 192.168.201.1 number = 0
2007 Sep 04
3
Ocfs2 and debian
Hi. I'm pretty new to ocfs2 and clusters. I'm trying to make ocfs2 running over a drbd device. I know it's not the best solution but for now i must deal with this. I set up drbd and work perfectly. I set up ocfs and i'm not able to make it to work. /etc/init.d/o2cb status: Module "configfs": Loaded Filesystem "configfs": Mounted Module
2008 Sep 10
4
mount.ocfs2: Error when attempting to run /sbin/ocfs2_hb_ctl: "Operation not permitted".
Hi, I am trying to configure a two node cluster on SLES10SP2 using user level heartbeat. Here is my configuration. ocfs2-tools-1.4.0-0.3 **user level heartbeat** -> lsmod | grep ocfs ocfs2_user_heartbeat 20992 1 ocfs2_dlmfs 37776 1 ocfs2_dlm 204456 1 ocfs2_dlmfs ocfs2_nodemanager 223384 6 ocfs2_user_heartbeat,ocfs2_dlmfs,ocfs2_dlm configfs 44700 3 ocfs2_user_heartbeat,ocfs2_nodemanager
2014 Aug 23
2
CTDB 50.samba: ERROR: smb.conf cache create failed
Ubuntu 14.04, ctdb 2.5.1 from the ubuntu package. samba 4.1.11 with cluster support from source. Error upon startup only. Restarting ctdb on that node clears the error. The other node with the same config is OK. Any ideas? Cheers, Steve smb.conf [global] workgroup = ALTEA realm = ALTEA.SITE security = ADS kerberos method = secrets and keytab netbios name = SMBCLUSTER disable netbios = Yes
2014 Jul 08
1
smbd does not start under ctdb
Hi 2 node drbd cluster with ocfs2. both nodes: openSUSE 4.1.9 with drbd 8.4 and ctdbd 2.3 All seems OK with ctdb: n1: ctdb status Number of nodes:2 pnn:0 192.168.0.10 OK (THIS NODE) pnn:1 192.168.0.11 OK Generation:1187222392 Size:2 hash:0 lmaster:0 hash:1 lmaster:1 Recovery mode:NORMAL (0) Recovery master:0 n2: ctdb status Number of nodes:2 pnn:0 192.168.0.10 OK pnn:1 192.168.0.11
2008 Mar 31
4
SuSe Hangs when /etc/init.d/o2cb online
Hello, I have a DELL MD3000i, and a couple of servers that I want to connect to the array. I have setup the main server with CentOS 5. - 2.6.18-53.el5 x86_64 - ocfs2-tools-1.2.7-1.el5 - ocfs2console-1.2.7-1.el5 - ocfs2-2.6.18-53.el5-1.2.8-2.el5 # Kernel Module And two OpenSuSe 10.3 Servers - 2.6.22.5-31-default x86_64 - ocfs2-tools-1.2.6-18 - ocfs2console-1.2.6-18 The module has been already
2011 Jun 27
1
multiple cluster doesn't work
We're trying to setup 3 PRDM partitions (VMware) across 2 nodes. As long as only one is configured in cluster.conf, there's not problem. As soon as we try to use 2 or more we get issues. It looks the same as bug 636: http://oss.oracle.com/bugzilla/show_bug.cgi?id=636 I posted my cluster.conf and command line results there. I'm including them here in the hopes that someone on this
2008 Sep 18
2
o2hb_do_disk_heartbeat:982:ERROR
Hi everyone; I have a problem on my 10 nodes cluster with ocfs2 1.2.9 and the OS is RHEL 4.7 AS. 9 nodes can start o2cb service and mount san disks on startup however one node can not do that. My cluster configuration is : node: ip_port = 7777 ip_address = 192.168.5.1 number = 0 name = fa01 cluster = ocfs2 node: ip_port =
2009 Oct 20
1
Fencing OCFS
I people, i install the ocfs in my virtual machine, with centos 5.3 and xen. But when i turn off the machine1, the ocfs start the fencing off the machine2. I read the doc in oracle.com, but could not solve the problem. Someone help? My conf and my package version. cluster.conf node: ip_port = 7777 ip_address = 192.168.1.35 number = 0 name = x1 cluster
2010 Sep 03
1
Servers reboot - may be OCFS2 related
Hello. What we have: 2x Debian 5.0 x64 - 2.6.32-20~bpo50+1 from backports DRBD + OCFS2 1.4.1-1 I have both node reboot every day on my tests. On heavy load it have to 1-3 hour to reboot, on idle about 20 hours. I not sure what it is a OCFS2 related but if I not mount OCFS2 partition - I don`t get this reboot. In attach screenshoot with system console on what error. Nothing special in logs.
2005 Jun 28
3
OCFS2 volumes do no mount automatically on RHEL4 also
On RHEL4 also, the service o2cb starts later than _netdev device mounts. So mount fails. Should the o2cb service start earlier to fix this? Zafar -----Original Message----- From: ocfs2-users-bounces@oss.oracle.com [mailto:ocfs2-users-bounces@oss.oracle.com] On Behalf Of ocfs2-users-request@oss.oracle.com Sent: Tuesday, June 28, 2005 12:00 PM To: ocfs2-users@oss.oracle.com Subject: Ocfs2-users
2005 Oct 12
2
Unable to access cluster service
hello, I'm running Ubuntu Breezy with the OCFS2 modules in the standard kernel. I installed ocfs2console and ocfs2-tools I've formatted a partition with ocfs2. But I can't add any node or mount the device(with the ocfs2console). because I get a "Unable to access cluster service" I can't find the cause nor the solution to this. root@lenaeja:~# /etc/init.d/o2cb status
2006 Mar 10
5
Add a new node to ocfs cluster
Hi everybody, My problem is that I want to add a new node to an existing RAC with ocfs2, without stop the database. If I add a new node to an existing ocfs cluster, do I need to stop the ocfs in the others nodes of the cluster? I've tried to do it without stop the ocfs in the others nodes but I get the following error: