search for: dlm_wait_for_lock_mastery

Displaying 15 results from an estimated 15 matches for "dlm_wait_for_lock_mastery".

2007 Nov 29
1
Troubles with two node
..._lock_resource:915 ERROR: status = -107 Nov 28 15:29:46 web-ha2 kernel: (23443,0):dlm_do_master_request:1331 ERROR: link to 0 went down! ERROR: status = -107 [...] Nov 22 18:14:50 web-ha2 kernel: (17634,0):dlm_restart_lock_mastery:1215 ERROR: node down! 0 Nov 22 18:14:50 web-ha2 kernel: (17634,0):dlm_wait_for_lock_mastery:1036 ERROR: status = -11 Nov 22 18:14:51 web-ha2 kernel: (17619,1):dlm_restart_lock_mastery:1215 ERROR: node down! 0 Nov 22 18:14:51 web-ha2 kernel: (17619,1):dlm_wait_for_lock_mastery:1036 ERROR: status = -11 Nov 22 18:14:51 web-ha2 kernel: (17798,1):dlm_restart_lock_mastery:1215 ERROR: node down!...
2023 Jun 16
1
[BUG] ocfs2/dlm: possible data races in dlm_drop_lockres_ref_done() and dlm_get_lock_resource()
...ged during the lockres lifecycle. So this won't cause any real problem since now it holds a reference. > > dlm_get_lock_resource() --> Line 701 in dlmmaster.c > if (res->owner != dlm->node_num) --> Line 1023 in dlmmaster.c (Access > res->owner) Do you mean in dlm_wait_for_lock_mastery()? Even if owner changes suddenly, it will recheck, so I think it is also fine. Thanks, Joseph > > The variables res->lockname.name and res->owner are accessed respectively > without holding the lock res->spinlock, and thus data races can occur. > > I am not quite sure w...
2007 Mar 08
4
ocfs2 cluster becomes unresponsive
...st master $RECOVERY lock now Mar 8 07:23:41 groupwise-1-mht kernel: (4432,0):ocfs2_replay_journal:1176 Recovering node 2 from slot 1 on device (253,1) Mar 8 07:23:41 groupwise-1-mht kernel: (4192,0):dlm_restart_lock_mastery:1214 ERROR: node down! 2 Mar 8 07:23:41 groupwise-1-mht kernel: (4192,0):dlm_wait_for_lock_mastery:1035 ERROR: status = -11 Mar 8 07:23:41 groupwise-1-mht kernel: (929,1):dlm_restart_lock_mastery:1214 ERROR: node down! 2 Mar 8 07:23:41 groupwise-1-mht kernel: (929,1):dlm_wait_for_lock_mastery:1035 ERROR: status = -11 Mar 8 07:23:42 groupwise-1-mht kernel: (4341,1):dlm_restart_lock_mastery:121...
2007 Jul 25
4
Problem installing on RH3 U8
Hi, i dont seem to be able to get ocfs running on RH3 U8 32Bit [root@libra-devb-db1 root]# uname -a Linux devb-db1.mydomain 2.4.21-47.ELsmp #1 SMP Wed Jul 5 20:38:41 EDT 2006 i686 athlon i386 GNU/Linux [root@devb-db1 root]# cat /etc/redhat-release Red Hat Enterprise Linux AS release 3 (Taroon Update 8) [root@devb-db1 root]# rpm -ivh ocfs-2.4.21-EL-smp-1.0.14-1.i686.rpm Preparing...
2014 Sep 11
1
May be deadlock for wrong locking order, patch request reviewed, thanks
...is held and the node did not release it which cause the cluster hangs up. root at cvknode-21:~# ps -e -o pid,stat,comm,wchan=WIDE-WCHAN-COLUMN | grep D PID STAT COMMAND WIDE-WCHAN-COLUMN 7489 D jbd2/sdh-621 jbd2_journal_commit_transaction 16218 D ls iterate_dir 16533 D mkdir dlm_wait_for_lock_mastery 31195 D+ ls iterate_dir So the code reviewed, and I found the order of the lock may wrong. In the function dlm_master_request_handler, the resource lock is held and so after the lock of &dlm->master_lock is locked. But in the function dlm_get_lock_resource, the &dlm->master_lock...
2014 Sep 11
1
May be deadlock for wrong locking order, patch request reviewed, thanks
...is held and the node did not release it which cause the cluster hangs up. root at cvknode-21:~# ps -e -o pid,stat,comm,wchan=WIDE-WCHAN-COLUMN | grep D PID STAT COMMAND WIDE-WCHAN-COLUMN 7489 D jbd2/sdh-621 jbd2_journal_commit_transaction 16218 D ls iterate_dir 16533 D mkdir dlm_wait_for_lock_mastery 31195 D+ ls iterate_dir So the code reviewed, and I found the order of the lock may wrong. In the function dlm_master_request_handler, the resource lock is held and so after the lock of &dlm->master_lock is locked. But in the function dlm_get_lock_resource, the &dlm->master_lock...
2010 Jan 14
1
another fencing question
Hi, periodically one of on my two nodes cluster is fenced here are the logs: Jan 14 07:01:44 nvr1-rc kernel: o2net: no longer connected to node nvr2- rc.minint.it (num 0) at 1.1.1.6:7777 Jan 14 07:01:44 nvr1-rc kernel: (21534,1):dlm_do_master_request:1334 ERROR: link to 0 went down! Jan 14 07:01:44 nvr1-rc kernel: (4007,4):dlm_send_proxy_ast_msg:458 ERROR: status = -112 Jan 14 07:01:44
2010 Apr 05
1
Kernel Panic, Server not coming back up
...o2net: accepted connection from node qa-web2 (num 2) at 147.178.220.32:7777 ocfs2_dlm: Node 2 joins domain 6A03E81A818641A68FD8DC23854E12D3 ocfs2_dlm: Nodes in domain ("6A03E81A818641A68FD8DC23854E12D3"): 0 1 2 (12701,1):dlm_restart_lock_mastery:1216 node 2 up while restarting (12701,1):dlm_wait_for_lock_mastery:1040 ERROR: status = -11 Any suggestions? Is there anymore data I can provide? Thanks for any help. Kevin
2023 Jun 13
1
[BUG] ocfs2/dlm: possible data races in dlm_drop_lockres_ref_done() and dlm_get_lock_resource()
Hello, Our static analysis tool finds some possible data races in the OCFS2 file system in Linux 6.4.0-rc6. In most calling contexts, the variables such as res->lockname.name and res->owner are accessed with holding the lock res->spinlock. Here is an example: lockres_seq_start() --> Line 539 in dlmdebug.c spin_lock(&res->spinlock); --> Line 574 in dlmdebug.c (Lock
2011 Dec 20
8
ocfs2 - Kernel panic on many write/read from both
Sorry i don`t copy everything: TEST-MAIL1# echo "ls //orphan_dir:0000"|debugfs.ocfs2 /dev/dm-0|wc debugfs.ocfs2 1.6.4 5239722 26198604 246266859 TEST-MAIL1# echo "ls //orphan_dir:0001"|debugfs.ocfs2 /dev/dm-0|wc debugfs.ocfs2 1.6.4 6074335 30371669 285493670 TEST-MAIL2 ~ # echo "ls //orphan_dir:0000"|debugfs.ocfs2 /dev/dm-0|wc debugfs.ocfs2 1.6.4 5239722 26198604
2010 Dec 09
2
servers blocked on ocfs2
...5.650496:1291450475.650501) Dec 4 09:15:06 parmenides kernel: o2net: no longer connected to node heraclito (num 0) at 192.168.1.3:7777 Dec 4 09:15:06 parmenides kernel: (snmpd,12342,11):dlm_do_master_request:1334 ERROR: link to 0 went down! Dec 4 09:15:06 parmenides kernel: (minilogd,12700,0):dlm_wait_for_lock_mastery:1117 ERROR: status = -112 Dec 4 09:15:06 parmenides kernel: (smbd,25555,12):dlm_do_master_request:1334 ERROR: link to 0 went down! Dec 4 09:15:06 parmenides kernel: (python,12439,9):dlm_do_master_request:1334 ERROR: link to 0 went down! Dec 4 09:15:06 parmenides kernel: (python,12439,9):dlm_g...
2013 Apr 28
2
Is it one issue. Do you have some good ideas, thanks a lot.
...D-VM6 kernel: [ 4231.992497] (dlm_reco_thread,14227,3):dlm_get_lock_resource:917 ERROR: status = -107 Apr 27 17:44:18 ZHJD-VM6 kernel: [ 4231.993204] (dlm_reco_thread,13736,2):dlm_restart_lock_mastery:1221 ERROR: node down! 2 Apr 27 17:44:18 ZHJD-VM6 kernel: [ 4231.993214] (dlm_reco_thread,13736,2):dlm_wait_for_lock_mastery:1038 ERROR: status = -11 Apr 27 17:44:18 ZHJD-VM6 kernel: [ 4231.993223] (dlm_reco_thread,13736,2):dlm_do_master_requery:1656 ERROR: Error -107 when sending message 514 (key 0xe00bcbbe) to node 3 Apr 27 17:44:18 ZHJD-VM6 kernel: [ 4231.993232] (dlm_reco_thread,13736,2):dlm_pre_master_reco_lockres:2...
2013 Apr 28
2
Is it one issue. Do you have some good ideas, thanks a lot.
...D-VM6 kernel: [ 4231.992497] (dlm_reco_thread,14227,3):dlm_get_lock_resource:917 ERROR: status = -107 Apr 27 17:44:18 ZHJD-VM6 kernel: [ 4231.993204] (dlm_reco_thread,13736,2):dlm_restart_lock_mastery:1221 ERROR: node down! 2 Apr 27 17:44:18 ZHJD-VM6 kernel: [ 4231.993214] (dlm_reco_thread,13736,2):dlm_wait_for_lock_mastery:1038 ERROR: status = -11 Apr 27 17:44:18 ZHJD-VM6 kernel: [ 4231.993223] (dlm_reco_thread,13736,2):dlm_do_master_requery:1656 ERROR: Error -107 when sending message 514 (key 0xe00bcbbe) to node 3 Apr 27 17:44:18 ZHJD-VM6 kernel: [ 4231.993232] (dlm_reco_thread,13736,2):dlm_pre_master_reco_lockres:2...
2007 Oct 08
2
OCF2 and LVM
Does anybody knows if is there a certified procedure in to backup a RAC DB 10.2.0.3 based on OCFS2 , via split mirror or snaphots technology ? Using Linux LVM and OCFS2, does anybody knows if is possible to dinamically extend an OCFS filesystem, once the underlying LVM Volume has been extended ? Thanks in advance Riccardo Paganini
2009 May 12
2
add error check for ocfs2_read_locked_inode() call
After upgrading from 2.6.28.10 to 2.6.29.3 I've saw following new errors in kernel log: May 12 14:46:41 falcon-cl5 May 12 14:46:41 falcon-cl5 (6757,7):ocfs2_read_locked_inode:466 ERROR: status = -22 Only one node is mounted volumes in cluster: /dev/sde on /home/apache/users/D1 type ocfs2 (rw,_netdev,noatime,heartbeat=local) /dev/sdd on /home/apache/users/D2 type ocfs2